Skip to content

Commit 619d45a

Browse files
committed
feat: new article
1 parent 9917f0c commit 619d45a

12 files changed

+256
-106
lines changed

_posts/Economics/2024-12-01-forecasting_commodity_prices_using_machine_learning_techniques_and_applications.md

Lines changed: 5 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -4,8 +4,7 @@ categories:
44
- Economics
55
classes: wide
66
date: '2024-12-01'
7-
excerpt: Explore how machine learning can be leveraged to forecast commodity prices,
8-
such as oil and gold, using advanced predictive models and economic indicators.
7+
excerpt: Explore how machine learning can be leveraged to forecast commodity prices, such as oil and gold, using advanced predictive models and economic indicators.
98
header:
109
image: /assets/images/data_science_13.jpg
1110
og_image: /assets/images/data_science_13.jpg
@@ -22,15 +21,11 @@ keywords:
2221
- Markdown
2322
- Data Science
2423
- Machine Learning
25-
seo_description: Learn how machine learning techniques are revolutionizing the forecasting
26-
of commodity prices like oil and gold, using advanced predictive models and economic
27-
indicators.
24+
- markdown
25+
seo_description: Learn how machine learning techniques are revolutionizing the forecasting of commodity prices like oil and gold, using advanced predictive models and economic indicators.
2826
seo_title: Forecasting Commodity Prices with Machine Learning | Data Science Applications
2927
seo_type: article
30-
summary: This article delves into the application of machine learning techniques to
31-
forecast commodity prices, such as oil and gold. It discusses the methods, economic
32-
indicators used, and the challenges in building predictive models in this complex
33-
domain.
28+
summary: This article delves into the application of machine learning techniques to forecast commodity prices, such as oil and gold. It discusses the methods, economic indicators used, and the challenges in building predictive models in this complex domain.
3429
tags:
3530
- Commodity prices
3631
- Machine learning
@@ -39,6 +34,7 @@ tags:
3934
- Data science in economics
4035
- Markdown
4136
- Data Science
37+
- markdown
4238
title: 'Forecasting Commodity Prices Using Machine Learning: Techniques and Applications'
4339
---
4440

_posts/data science/2019-12-29-understanding_splines_what_they_how_they_used_data_analysis.md

Lines changed: 3 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -4,9 +4,7 @@ categories:
44
- Data Science
55
classes: wide
66
date: '2019-12-29'
7-
excerpt: Splines are powerful tools for modeling complex, nonlinear relationships
8-
in data. In this article, we'll explore what splines are, how they work, and how
9-
they are used in data analysis, statistics, and machine learning.
7+
excerpt: Splines are powerful tools for modeling complex, nonlinear relationships in data. In this article, we'll explore what splines are, how they work, and how they are used in data analysis, statistics, and machine learning.
108
header:
119
image: /assets/images/data_science_19.jpg
1210
og_image: /assets/images/data_science_19.jpg
@@ -25,15 +23,10 @@ keywords:
2523
- Go
2624
- Statistics
2725
- Machine Learning
28-
seo_description: Splines are flexible mathematical tools used for smoothing and modeling
29-
complex data patterns. Learn what they are, how they work, and their practical applications
30-
in regression, data smoothing, and machine learning.
26+
seo_description: Splines are flexible mathematical tools used for smoothing and modeling complex data patterns. Learn what they are, how they work, and their practical applications in regression, data smoothing, and machine learning.
3127
seo_title: What Are Splines? A Deep Dive into Their Uses in Data Analysis
3228
seo_type: article
33-
summary: Splines are flexible mathematical functions used to approximate complex patterns
34-
in data. They help smooth data, model non-linear relationships, and fit curves in
35-
regression analysis. This article covers the basics of splines, their various types,
36-
and their practical applications in statistics, data science, and machine learning.
29+
summary: Splines are flexible mathematical functions used to approximate complex patterns in data. They help smooth data, model non-linear relationships, and fit curves in regression analysis. This article covers the basics of splines, their various types, and their practical applications in statistics, data science, and machine learning.
3730
tags:
3831
- Splines
3932
- Regression

_posts/data science/2019-12-30-evaluating_binary_classifiers_imbalanced_datasets.md

Lines changed: 4 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -4,9 +4,7 @@ categories:
44
- Data Science
55
classes: wide
66
date: '2019-12-30'
7-
excerpt: AUC-ROC and Gini are popular metrics for evaluating binary classifiers, but
8-
they can be misleading on imbalanced datasets. Discover why AUC-PR, with its focus
9-
on Precision and Recall, offers a better evaluation for handling rare events.
7+
excerpt: AUC-ROC and Gini are popular metrics for evaluating binary classifiers, but they can be misleading on imbalanced datasets. Discover why AUC-PR, with its focus on Precision and Recall, offers a better evaluation for handling rare events.
108
header:
119
image: /assets/images/data_science_8.jpg
1210
og_image: /assets/images/data_science_8.jpg
@@ -20,23 +18,16 @@ keywords:
2018
- Binary classifiers
2119
- Imbalanced data
2220
- Machine learning metrics
23-
seo_description: When evaluating binary classifiers on imbalanced datasets, AUC-PR
24-
is a more informative metric than AUC-ROC or Gini. Learn why Precision-Recall curves
25-
provide a clearer picture of model performance on rare events.
21+
seo_description: When evaluating binary classifiers on imbalanced datasets, AUC-PR is a more informative metric than AUC-ROC or Gini. Learn why Precision-Recall curves provide a clearer picture of model performance on rare events.
2622
seo_title: 'AUC-PR vs. AUC-ROC: Evaluating Classifiers on Imbalanced Data'
2723
seo_type: article
28-
summary: In this article, we explore why AUC-PR (Area Under Precision-Recall Curve)
29-
is a superior metric for evaluating binary classifiers on imbalanced datasets compared
30-
to AUC-ROC and Gini. We discuss how class imbalance distorts performance metrics
31-
and provide real-world examples of why Precision-Recall curves give a clearer understanding
32-
of model performance on rare events.
24+
summary: In this article, we explore why AUC-PR (Area Under Precision-Recall Curve) is a superior metric for evaluating binary classifiers on imbalanced datasets compared to AUC-ROC and Gini. We discuss how class imbalance distorts performance metrics and provide real-world examples of why Precision-Recall curves give a clearer understanding of model performance on rare events.
3325
tags:
3426
- Binary classifiers
3527
- Imbalanced data
3628
- Auc-pr
3729
- Precision-recall
38-
title: 'Evaluating Binary Classifiers on Imbalanced Datasets: Why AUC-PR Beats AUC-ROC
39-
and Gini'
30+
title: 'Evaluating Binary Classifiers on Imbalanced Datasets: Why AUC-PR Beats AUC-ROC and Gini'
4031
---
4132

4233
When working with binary classifiers, metrics like **AUC-ROC** and **Gini** have long been the default for evaluating model performance. These metrics offer a quick way to assess how well a model discriminates between two classes, typically a **positive class** (e.g., detecting fraud or predicting defaults) and a **negative class** (e.g., non-fraudulent or non-default cases).

_posts/data science/2020-01-06-role_data_science_predictive_maintenance.md

Lines changed: 3 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -4,9 +4,7 @@ categories:
44
- Data Science
55
classes: wide
66
date: '2020-01-06'
7-
excerpt: Explore the role of data science in predictive maintenance, from forecasting
8-
equipment failure to optimizing maintenance schedules using techniques like regression
9-
and anomaly detection.
7+
excerpt: Explore the role of data science in predictive maintenance, from forecasting equipment failure to optimizing maintenance schedules using techniques like regression and anomaly detection.
108
header:
119
image: /assets/images/data_science_7.jpg
1210
og_image: /assets/images/data_science_7.jpg
@@ -21,14 +19,10 @@ keywords:
2119
- Machine learning
2220
- Predictive analytics
2321
- Industrial analytics
24-
seo_description: Discover how data science techniques such as regression, clustering,
25-
and anomaly detection optimize predictive maintenance, helping organizations forecast
26-
failures and enhance operational efficiency.
22+
seo_description: Discover how data science techniques such as regression, clustering, and anomaly detection optimize predictive maintenance, helping organizations forecast failures and enhance operational efficiency.
2723
seo_title: How Data Science Powers Predictive Maintenance
2824
seo_type: article
29-
summary: An in-depth look at how data science techniques such as regression, clustering,
30-
anomaly detection, and machine learning are transforming predictive maintenance
31-
across various industries.
25+
summary: An in-depth look at how data science techniques such as regression, clustering, anomaly detection, and machine learning are transforming predictive maintenance across various industries.
3226
tags:
3327
- Predictive maintenance
3428
- Machine learning

_posts/statistics/2019-12-28-shapirowilk_test_vs_andersondarling_checking_normality_small_large_samples.md

Lines changed: 7 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -4,9 +4,7 @@ categories:
44
- Statistics
55
classes: wide
66
date: '2019-12-28'
7-
excerpt: Explore the differences between the Shapiro-Wilk and Anderson-Darling tests,
8-
two common methods for testing normality, and how sample size and distribution affect
9-
their performance.
7+
excerpt: Explore the differences between the Shapiro-Wilk and Anderson-Darling tests, two common methods for testing normality, and how sample size and distribution affect their performance.
108
header:
119
image: /assets/images/data_science_20.jpg
1210
og_image: /assets/images/data_science_20.jpg
@@ -22,22 +20,19 @@ keywords:
2220
- Large sample size
2321
- Statistical distribution
2422
- Python
25-
seo_description: A comparison of the Shapiro-Wilk and Anderson-Darling tests for normality,
26-
analyzing their strengths and weaknesses based on sample size and distribution.
27-
seo_title: 'Shapiro-Wilk vs Anderson-Darling: Normality Tests for Small and Large
28-
Samples'
23+
- python
24+
seo_description: A comparison of the Shapiro-Wilk and Anderson-Darling tests for normality, analyzing their strengths and weaknesses based on sample size and distribution.
25+
seo_title: 'Shapiro-Wilk vs Anderson-Darling: Normality Tests for Small and Large Samples'
2926
seo_type: article
30-
summary: This article compares the Shapiro-Wilk and Anderson-Darling tests, emphasizing
31-
how sample size and distribution characteristics influence the choice of method
32-
when assessing normality.
27+
summary: This article compares the Shapiro-Wilk and Anderson-Darling tests, emphasizing how sample size and distribution characteristics influence the choice of method when assessing normality.
3328
tags:
3429
- Normality testing
3530
- Shapiro-wilk test
3631
- Anderson-darling test
3732
- Sample size
3833
- Python
39-
title: 'Shapiro-Wilk Test vs. Anderson-Darling: Checking for Normality in Small vs.
40-
Large Samples'
34+
- python
35+
title: 'Shapiro-Wilk Test vs. Anderson-Darling: Checking for Normality in Small vs. Large Samples'
4136
---
4237

4338
## Shapiro-Wilk Test vs. Anderson-Darling: Checking for Normality in Small vs. Large Samples

_posts/statistics/2019-12-31-deep_dive_into_why_multiple_imputation_indefensible.md

Lines changed: 3 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -4,8 +4,7 @@ categories:
44
- Statistics
55
classes: wide
66
date: '2019-12-31'
7-
excerpt: Let's examine why multiple imputation, despite being popular, may not be
8-
as robust or interpretable as it's often considered. Is there a better approach?
7+
excerpt: Let's examine why multiple imputation, despite being popular, may not be as robust or interpretable as it's often considered. Is there a better approach?
98
header:
109
image: /assets/images/data_science_20.jpg
1110
og_image: /assets/images/data_science_20.jpg
@@ -18,14 +17,10 @@ keywords:
1817
- Missing data
1918
- Single stochastic imputation
2019
- Deterministic sensitivity analysis
21-
seo_description: Exploring the issues with multiple imputation and why single stochastic
22-
imputation with deterministic sensitivity analysis is a superior alternative.
20+
seo_description: Exploring the issues with multiple imputation and why single stochastic imputation with deterministic sensitivity analysis is a superior alternative.
2321
seo_title: 'The Case Against Multiple Imputation: An In-depth Look'
2422
seo_type: article
25-
summary: Multiple imputation is widely regarded as the gold standard for handling
26-
missing data, but it carries significant conceptual and interpretative challenges.
27-
We will explore its weaknesses and propose an alternative using single stochastic
28-
imputation and deterministic sensitivity analysis.
23+
summary: Multiple imputation is widely regarded as the gold standard for handling missing data, but it carries significant conceptual and interpretative challenges. We will explore its weaknesses and propose an alternative using single stochastic imputation and deterministic sensitivity analysis.
2924
tags:
3025
- Multiple imputation
3126
- Missing data

_posts/statistics/2020-01-01-causality_correlation.md

Lines changed: 3 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -4,8 +4,7 @@ categories:
44
- Statistics
55
classes: wide
66
date: '2020-01-01'
7-
excerpt: Understand how causal reasoning helps us move beyond correlation, resolving
8-
paradoxes and leading to more accurate insights from data analysis.
7+
excerpt: Understand how causal reasoning helps us move beyond correlation, resolving paradoxes and leading to more accurate insights from data analysis.
98
header:
109
image: /assets/images/data_science_4.jpg
1110
og_image: /assets/images/data_science_1.jpg
@@ -19,14 +18,10 @@ keywords:
1918
- Berkson's paradox
2019
- Correlation
2120
- Data science
22-
seo_description: Explore how causal reasoning, through paradoxes like Simpson's and
23-
Berkson's, can help us avoid the common pitfalls of interpreting data solely based
24-
on correlation.
21+
seo_description: Explore how causal reasoning, through paradoxes like Simpson's and Berkson's, can help us avoid the common pitfalls of interpreting data solely based on correlation.
2522
seo_title: 'Causality Beyond Correlation: Understanding Paradoxes and Causal Graphs'
2623
seo_type: article
27-
summary: An in-depth exploration of the limits of correlation in data interpretation,
28-
highlighting Simpson's and Berkson's paradoxes and introducing causal graphs as
29-
a tool for uncovering true causal relationships.
24+
summary: An in-depth exploration of the limits of correlation in data interpretation, highlighting Simpson's and Berkson's paradoxes and introducing causal graphs as a tool for uncovering true causal relationships.
3025
tags:
3126
- Simpson's paradox
3227
- Berkson's paradox

_posts/statistics/2020-01-02-maximum_likelihood_estimation_statistical_modeling.md

Lines changed: 5 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -4,9 +4,7 @@ categories:
44
- Statistics
55
classes: wide
66
date: '2020-01-02'
7-
excerpt: Discover the fundamentals of Maximum Likelihood Estimation (MLE), its role
8-
in data science, and how it impacts businesses through predictive analytics and
9-
risk modeling.
7+
excerpt: Discover the fundamentals of Maximum Likelihood Estimation (MLE), its role in data science, and how it impacts businesses through predictive analytics and risk modeling.
108
header:
119
image: /assets/images/data_science_3.jpg
1210
og_image: /assets/images/data_science_3.jpg
@@ -22,20 +20,19 @@ keywords:
2220
- Mle
2321
- Bash
2422
- Python
25-
seo_description: Explore Maximum Likelihood Estimation (MLE), its importance in data
26-
science, machine learning, and real-world applications.
23+
seo_description: Explore Maximum Likelihood Estimation (MLE), its importance in data science, machine learning, and real-world applications.
2724
seo_title: 'MLE: A Key Tool in Data Science'
2825
seo_type: article
29-
summary: This article covers the essentials of Maximum Likelihood Estimation (MLE),
30-
breaking down its mathematical foundation, importance in data science, practical
31-
applications, and limitations.
26+
summary: This article covers the essentials of Maximum Likelihood Estimation (MLE), breaking down its mathematical foundation, importance in data science, practical applications, and limitations.
3227
tags:
3328
- Statistical modeling
3429
- Bash
3530
- Maximum likelihood estimation
3631
- Data science
3732
- Mle
3833
- Python
34+
- python
35+
- bash
3936
title: 'Maximum Likelihood Estimation (MLE): Statistical Modeling in Data Science'
4037
---
4138

_posts/statistics/2020-01-03-assessing_goodnessoffit_nonparametric_data.md

Lines changed: 3 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -4,9 +4,7 @@ categories:
44
- Statistics
55
classes: wide
66
date: '2020-01-03'
7-
excerpt: The Kolmogorov-Smirnov test is a powerful tool for assessing goodness-of-fit
8-
in non-parametric data. Learn how it works, how it compares to the Shapiro-Wilk
9-
test, and explore real-world applications.
7+
excerpt: The Kolmogorov-Smirnov test is a powerful tool for assessing goodness-of-fit in non-parametric data. Learn how it works, how it compares to the Shapiro-Wilk test, and explore real-world applications.
108
header:
119
image: /assets/images/data_science_3.jpg
1210
og_image: /assets/images/data_science_3.jpg
@@ -20,15 +18,10 @@ keywords:
2018
- Non-parametric statistics
2119
- Distribution fitting
2220
- Shapiro-wilk test
23-
seo_description: This article introduces the Kolmogorov-Smirnov test for assessing
24-
goodness-of-fit in non-parametric data, comparing it with other tests like Shapiro-Wilk,
25-
and exploring real-world use cases.
21+
seo_description: This article introduces the Kolmogorov-Smirnov test for assessing goodness-of-fit in non-parametric data, comparing it with other tests like Shapiro-Wilk, and exploring real-world use cases.
2622
seo_title: 'Kolmogorov-Smirnov Test: A Guide to Non-Parametric Goodness-of-Fit Testing'
2723
seo_type: article
28-
summary: This article explains the Kolmogorov-Smirnov (K-S) test for assessing the
29-
goodness-of-fit of non-parametric data. We compare the K-S test to other goodness-of-fit
30-
tests, such as Shapiro-Wilk, and provide real-world use cases, including testing
31-
whether a dataset follows a specific distribution.
24+
summary: This article explains the Kolmogorov-Smirnov (K-S) test for assessing the goodness-of-fit of non-parametric data. We compare the K-S test to other goodness-of-fit tests, such as Shapiro-Wilk, and provide real-world use cases, including testing whether a dataset follows a specific distribution.
3225
tags:
3326
- Kolmogorov-smirnov test
3427
- Goodness-of-fit tests

_posts/statistics/2020-01-04-multiple_comparisons_problem_bonferroni_correction_other_solutions.md

Lines changed: 3 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -4,9 +4,7 @@ categories:
44
- Statistics
55
classes: wide
66
date: '2020-01-04'
7-
excerpt: The multiple comparisons problem arises in hypothesis testing when performing
8-
multiple tests increases the likelihood of false positives. Learn about the Bonferroni
9-
correction and other solutions to control error rates.
7+
excerpt: The multiple comparisons problem arises in hypothesis testing when performing multiple tests increases the likelihood of false positives. Learn about the Bonferroni correction and other solutions to control error rates.
108
header:
119
image: /assets/images/data_science_6.jpg
1210
og_image: /assets/images/data_science_6.jpg
@@ -21,15 +19,10 @@ keywords:
2119
- False discovery rate
2220
- Hypothesis testing
2321
- Python
24-
seo_description: This article explains the multiple comparisons problem in hypothesis
25-
testing and discusses solutions such as Bonferroni correction, Holm-Bonferroni,
26-
and FDR, with practical applications in fields like medical studies and genetics.
22+
seo_description: This article explains the multiple comparisons problem in hypothesis testing and discusses solutions such as Bonferroni correction, Holm-Bonferroni, and FDR, with practical applications in fields like medical studies and genetics.
2723
seo_title: 'Understanding the Multiple Comparisons Problem: Bonferroni and Other Solutions'
2824
seo_type: article
29-
summary: This article explores the multiple comparisons problem in hypothesis testing,
30-
discussing solutions like the Bonferroni correction, Holm-Bonferroni method, and
31-
False Discovery Rate (FDR). It includes practical examples from experiments involving
32-
multiple testing, such as medical studies and genetics.
25+
summary: This article explores the multiple comparisons problem in hypothesis testing, discussing solutions like the Bonferroni correction, Holm-Bonferroni method, and False Discovery Rate (FDR). It includes practical examples from experiments involving multiple testing, such as medical studies and genetics.
3326
tags:
3427
- Multiple comparisons problem
3528
- Bonferroni correction

0 commit comments

Comments
 (0)