Skip to content

Commit 9df6c38

Browse files
Merge pull request #88 from DiogoRibeiro7/feat/reserve_branche
Feat/reserve branche
2 parents 7daa93e + 6b6aa40 commit 9df6c38

15 files changed

+667
-42
lines changed

_posts/-_ideas/2030-01-01-ideas_statistical_tests.md

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -36,11 +36,6 @@ title: 15 Article Ideas on Statistical Tests
3636

3737
Here are some interesting article ideas centered around statistical tests, designed to help you explore the differences, applications, and nuances of various statistical methods:
3838

39-
### 1. **"T-Test vs. Z-Test: When and Why to Use Each"**
40-
- Explain the differences between the t-test and z-test.
41-
- Discuss when each test is appropriate based on sample size, variance, and distribution.
42-
- Provide real-world applications for each test.
43-
- Explore one-sample, two-sample, and paired t-tests.
4439

4540
### 2. **"Chi-Square Test: Applications in Categorical Data Analysis"**
4641
- Overview of the chi-square test for independence and goodness of fit.

_posts/-_ideas/2030-01-01-new_articles_topics.md

Lines changed: 1 addition & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -25,13 +25,8 @@ There are several interesting article topics you can explore under the umbrella
2525
- **Overview**: An in-depth guide on how machine learning models are applied in PdM, covering supervised, unsupervised, and reinforcement learning techniques.
2626
- **Focus**: How models like decision trees, random forests, support vector machines (SVM), and neural networks are used to predict equipment failures.
2727

28-
### 5. A Comparison of Predictive Maintenance Algorithms: Classical vs. Machine Learning Approaches
29-
- **Overview**: Compare traditional predictive maintenance models (e.g., statistical models like ARIMA) with machine learning algorithms.
30-
- **Focus**: Performance, accuracy, and scalability differences between classical approaches and machine learning models in real-world applications.
3128

32-
### 6. IoT and Sensor Data: The Backbone of Predictive Maintenance
33-
- **Overview**: Explain how IoT-enabled devices and sensors gather data that drives predictive maintenance strategies.
34-
- **Focus**: Types of sensors (vibration, temperature, pressure, etc.), the importance of real-time monitoring, and how this data is utilized for predictive maintenance.
29+
3530

3631
### 7. Deep Learning for Predictive Maintenance: Unlocking Hidden Patterns in Data
3732
- **Overview**: Explore how deep learning models such as convolutional neural networks (CNN) and recurrent neural networks (RNN) are used for complex PdM scenarios.

_posts/2020-01-06-role_data_science_predictive_maintenance.md

Lines changed: 0 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -32,22 +32,6 @@ tags:
3232
title: Leveraging Data Science Techniques for Predictive Maintenance
3333
---
3434

35-
## Table of Contents
36-
37-
1. Introduction to Predictive Maintenance (PdM)
38-
2. The Importance of Data Science in PdM
39-
3. Key Data Science Techniques in Predictive Maintenance
40-
1. Regression Analysis
41-
2. Anomaly Detection
42-
3. Clustering Algorithms
43-
4. Data Requirements and Challenges in PdM
44-
5. Role of Machine Learning in Predictive Maintenance
45-
6. Applications of PdM Across Industries
46-
7. Future of Data Science in Predictive Maintenance
47-
8. Conclusion
48-
49-
---
50-
5135
## 1. Introduction to Predictive Maintenance (PdM)
5236

5337
Predictive maintenance (PdM) refers to the practice of using data-driven techniques to predict when equipment will fail, allowing for timely and efficient maintenance. This proactive approach aims to reduce downtime, optimize equipment lifespan, and minimize maintenance costs. Unlike traditional maintenance strategies, such as reactive (fixing after failure) or preventive (servicing at regular intervals), PdM leverages real-time data, statistical analysis, and predictive models to forecast equipment degradation and identify the optimal time for intervention.

_posts/2021-05-10-estimating_uncertainty_neural_networks_using_monte_carlo_dropout.md

Lines changed: 9 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -54,7 +54,7 @@ Dropout is a regularization technique commonly used to prevent overfitting in ne
5454

5555
Monte Carlo dropout, introduced by Yarin Gal and his colleagues, builds on this technique by keeping dropout enabled during inference. This seemingly simple modification allows the model to behave like a Bayesian approximation, enabling it to produce a distribution of outputs for a given input. By running the neural network multiple times on the same input (with different dropout masks applied each time), we can approximate the posterior predictive distribution of the model’s outputs.
5656

57-
Mathematically, if $f(y|x)$ denotes the output of the neural network for class $y$ on input $x$, then the Monte Carlo dropout approach involves drawing multiple samples from $f(y|x)$ by running the model several times with dropout enabled. These samples can be used to compute the mean and variance of the model's predictions, which serve as estimates of the predictive mean $\mathbb{E}[f(y|x)]$ and predictive variance $\text{Var}[f(y|x)]$.
57+
Mathematically, if $$f(y|x)$$ denotes the output of the neural network for class $$y$$ on input $$x$$, then the Monte Carlo dropout approach involves drawing multiple samples from $$f(y|x)$$ by running the model several times with dropout enabled. These samples can be used to compute the mean and variance of the model's predictions, which serve as estimates of the predictive mean $$\mathbb{E}[f(y|x)]$$ and predictive variance $$\text{Var}[f(y|x)]$$.
5858

5959
This technique provides a straightforward way to quantify the uncertainty of a model's predictions. In practice, Monte Carlo dropout is used to estimate uncertainty in both classification and regression tasks, although our focus here will be on multi-class classification.
6060

@@ -72,13 +72,13 @@ Monte Carlo dropout works by approximating the posterior distribution of a model
7272

7373
### Formalizing the Process
7474

75-
Let $f(y|x)$ be the softmax output of the neural network for class $y$ given input $x$. Monte Carlo dropout involves generating $T$ samples $\{ f_t(y|x) \}_{t=1}^{T}$ by running the network $T$ times with different dropout masks. From these samples, we can compute:
75+
Let $$f(y|x)$$ be the softmax output of the neural network for class $$y$$ given input $$x$$. Monte Carlo dropout involves generating $$T$$ samples $$\{ f_t(y|x) \}_{t=1}^{T}$$ by running the network $$T$$ times with different dropout masks. From these samples, we can compute:
7676

7777
- **Predictive mean**:
7878
$$
7979
\mathbb{E}[f(y|x)] = \frac{1}{T} \sum_{t=1}^{T} f_t(y|x)
8080
$$
81-
This gives the average probability assigned to class $y$ across the $T$ stochastic forward passes.
81+
This gives the average probability assigned to class $$y$$ across the $$T$$ stochastic forward passes.
8282

8383
- **Predictive variance**:
8484
$$
@@ -100,7 +100,7 @@ $$
100100
\text{Uncertainty Score} = 1 - \max_y \mathbb{E}[f(y|x)]
101101
$$
102102

103-
This score measures the model's confidence in its most likely prediction. A high value for $\max_y \mathbb{E}[f(y|x)]$ indicates high confidence in the predicted class, while a lower value suggests greater uncertainty.
103+
This score measures the model's confidence in its most likely prediction. A high value for $$\max_y \mathbb{E}[f(y|x)]$$ indicates high confidence in the predicted class, while a lower value suggests greater uncertainty.
104104

105105
This method is simple and easy to implement, but it has some limitations. For example, it only takes into account the predicted class's probability and ignores the spread of probabilities across other classes. In cases where the model assigns similar probabilities to multiple classes, this method might underestimate uncertainty.
106106

@@ -118,7 +118,7 @@ This method captures uncertainty more comprehensively than the maximum class pro
118118

119119
### 3. Variance-Based Uncertainty Estimation
120120

121-
Another method is to use the variance of the predicted probabilities as a measure of uncertainty. The variance for each class $y$ is computed as:
121+
Another method is to use the variance of the predicted probabilities as a measure of uncertainty. The variance for each class $$y$$ is computed as:
122122

123123
$$
124124
\text{Var}[f(y|x)] = \frac{1}{T} \sum_{t=1}^{T} (f_t(y|x) - \mathbb{E}[f(y|x)])^2
@@ -136,19 +136,19 @@ Variance-based methods are particularly useful when the goal is to detect out-of
136136

137137
### 4. Error Function and Normal Approximation
138138

139-
In some cases, particularly when dealing with binary or reduced two-class problems, it may be useful to approximate the predictive distribution using a normal distribution. Specifically, we can model the output probabilities for class $y$ as a Gaussian distribution:
139+
In some cases, particularly when dealing with binary or reduced two-class problems, it may be useful to approximate the predictive distribution using a normal distribution. Specifically, we can model the output probabilities for class $$y$$ as a Gaussian distribution:
140140

141141
$$
142142
p(y|x) \sim \mathcal{N}(\mu_y, \sigma_y^2)
143143
$$
144-
where $\mu_y = \mathbb{E}[f(y|x)]$ is the predictive mean and $\sigma_y^2 = \text{Var}[f(y|x)]$ is the predictive variance.
144+
where $$\mu_y = \mathbb{E}[f(y|x)]$$ is the predictive mean and $$\sigma_y^2 = \text{Var}[f(y|x)]$$ is the predictive variance.
145145

146-
For a two-class classifier, let $y$ be the predicted class (i.e., $y = \arg\max_y \mathbb{E}[f(y|x)]$) and $\neg y$ be the other class. The probability that a future evaluation of the classifier will also output $y$ is given by:
146+
For a two-class classifier, let $$y$$ be the predicted class (i.e., $$y = \arg\max_y \mathbb{E}[f(y|x)]$$) and $$\neg y$$ be the other class. The probability that a future evaluation of the classifier will also output $$y$$ is given by:
147147

148148
$$
149149
u = \Pr[X \geq 0]
150150
$$
151-
where $X \sim \mathcal{N}(\mu_y - \mu_{\neg y}, \sigma_y^2 + \sigma_{\neg y}^2)$.
151+
where $$X \sim \mathcal{N}(\mu_y - \mu_{\neg y}, \sigma_y^2 + \sigma_{\neg y}^2)$$.
152152

153153
This probability can be estimated using the error function:
154154

0 commit comments

Comments
 (0)