Skip to content

Commit e6deef7

Browse files
committed
Fix typos in the doc
1 parent f85f12c commit e6deef7

13 files changed

+18
-18
lines changed

docs/examples/batch-to-online.ipynb

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -60,7 +60,7 @@
6060
" ('lin_reg', linear_model.LogisticRegression(solver='lbfgs'))\n",
6161
"])\n",
6262
"\n",
63-
"# Define a determistic cross-validation procedure\n",
63+
"# Define a deterministic cross-validation procedure\n",
6464
"cv = model_selection.KFold(n_splits=5, shuffle=True, random_state=42)\n",
6565
"\n",
6666
"# Compute the MSE values\n",
@@ -356,7 +356,7 @@
356356
"cell_type": "markdown",
357357
"metadata": {},
358358
"source": [
359-
"The results seem to be exactly the same! The twist is that the running statistics won't be very accurate for the first few observations. In general though this doesn't matter too much. Some would even go as far as to say that this descrepancy is beneficial and acts as some sort of regularization...\n",
359+
"The results seem to be exactly the same! The twist is that the running statistics won't be very accurate for the first few observations. In general though this doesn't matter too much. Some would even go as far as to say that this discrepancy is beneficial and acts as some sort of regularization...\n",
360360
"\n",
361361
"Now the idea is that we can compute the running statistics of each feature and scale them as they come along. The way to do this with River is to use the `StandardScaler` class from the `preprocessing` module, as so:"
362362
]

docs/examples/building-a-simple-nowcasting-model.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -446,7 +446,7 @@
446446
"cell_type": "markdown",
447447
"metadata": {},
448448
"source": [
449-
"We've managed to get a good looking prediction curve with a reasonably simple model. What's more our model has the advantage of being interpretable and easy to debug. There surely are more rocks to squeeze (e.g. tune the hyperparameters, use an ensemble model, etc.) but we'll leave that as an exercice to the reader.\n",
449+
"We've managed to get a good looking prediction curve with a reasonably simple model. What's more our model has the advantage of being interpretable and easy to debug. There surely are more rocks to squeeze (e.g. tune the hyperparameters, use an ensemble model, etc.) but we'll leave that as an exercise to the reader.\n",
450450
"\n",
451451
"As a finishing touch we'll rewrite our pipeline using the `|` operator, which is called a \"pipe\"."
452452
]

docs/examples/content-personalization.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -319,7 +319,7 @@
319319
"cell_type": "markdown",
320320
"metadata": {},
321321
"source": [
322-
"A good recommender model should at the very least understand what kind of items each user prefers. One of the simplest and yet performant way to do this is Simon Funk's SGD method he developped for the Netflix challenge and wrote about [here](https://sifter.org/simon/journal/20061211.html). It models each user and each item as latent vectors. The dot product of these two vectors is the expected preference of the user for the item."
322+
"A good recommender model should at the very least understand what kind of items each user prefers. One of the simplest and yet performant way to do this is Simon Funk's SGD method he developed for the Netflix challenge and wrote about [here](https://sifter.org/simon/journal/20061211.html). It models each user and each item as latent vectors. The dot product of these two vectors is the expected preference of the user for the item."
323323
]
324324
},
325325
{

docs/examples/sentence-classification.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -814,7 +814,7 @@
814814
"cell_type": "markdown",
815815
"metadata": {},
816816
"source": [
817-
"The command below allows you to download the pre-trained embeddings that spaCy makes available. More informations about spaCy and its installation may be found here [here](https://spacy.io/usage)."
817+
"The command below allows you to download the pre-trained embeddings that spaCy makes available. More information about spaCy and its installation may be found here [here](https://spacy.io/usage)."
818818
]
819819
},
820820
{

docs/faq/index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -58,4 +58,4 @@ There are many great open-source libraries for building neural network models. W
5858

5959
## Who are the authors of this library?
6060

61-
We are research engineers, graduate students, PhDs and machine learning researchers. The members of the develompent team are mainly located in France, Brazil and New Zealand.
61+
We are research engineers, graduate students, PhDs and machine learning researchers. The members of the development team are mainly located in France, Brazil and New Zealand.

docs/introduction/basic-concepts.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,7 @@ Dictionaries are therefore a perfect fit. They're native to Python and have exce
4444

4545
In production, you're almost always going to face data streams which you have to react to, such as users visiting your website. The advantage of online machine learning is that you can design models that make predictions as well as learn from this data stream as it flows.
4646

47-
But of course, when you're developping a model, you don't usually have access to a real-time feed on which to evaluate your model. You usually have an offline dataset which you want to evaluate your model on. River provides some datasets which can be read in online manner, one sample at a time. It is however crucial to keep in mind that the goal is to reproduce a production scenario as closely as possible, in order to ensure your model will perform just as well in production.
47+
But of course, when you're developing a model, you don't usually have access to a real-time feed on which to evaluate your model. You usually have an offline dataset which you want to evaluate your model on. River provides some datasets which can be read in online manner, one sample at a time. It is however crucial to keep in mind that the goal is to reproduce a production scenario as closely as possible, in order to ensure your model will perform just as well in production.
4848

4949
## Model evaluation
5050

docs/introduction/getting-started/concept-drift-detection.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -179,7 +179,7 @@
179179
}
180180
},
181181
"source": [
182-
"We see that `ADWIN` successfully indicates the presence of drift (red vertical lines) close to the begining of a new data distribution.\n",
182+
"We see that `ADWIN` successfully indicates the presence of drift (red vertical lines) close to the beginning of a new data distribution.\n",
183183
"\n",
184184
"\n",
185185
"---\n",

docs/recipes/active-learning.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -196,7 +196,7 @@
196196
"cell_type": "markdown",
197197
"metadata": {},
198198
"source": [
199-
"Active learning is primarly used to label data in an efficient manner. However, in an online setting, active learning can also be used simply to speed up training. The point is that you can achieve a very good performance without training on an entire dataset. Active learning is a powerful way to decide which samples to train on."
199+
"Active learning is primarily used to label data in an efficient manner. However, in an online setting, active learning can also be used simply to speed up training. The point is that you can achieve a very good performance without training on an entire dataset. Active learning is a powerful way to decide which samples to train on."
200200
]
201201
},
202202
{

docs/recipes/cloning-and-mutating.ipynb

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@
1313
"source": [
1414
"Sometimes you might want to reset a model, or edit (what we call mutate) its attributes. This can be useful in an online environment. Indeed, if you detect a drift, then you might want to mutate a model's attributes. Or if you see that a model's performance is plummeting, then you might to reset it to its \"factory settings\".\n",
1515
"\n",
16-
"Anyway, this is not to convince you, but rather to say that a model's attributes don't have be to set in stone throughout its lifetime. In particular, if you're developping your own model, then you might want to have good tools to do this. This is what this recipe is about."
16+
"Anyway, this is not to convince you, but rather to say that a model's attributes don't have be to set in stone throughout its lifetime. In particular, if you're developing your own model, then you might want to have good tools to do this. This is what this recipe is about."
1717
]
1818
},
1919
{
@@ -332,9 +332,9 @@
332332
"cell_type": "markdown",
333333
"metadata": {},
334334
"source": [
335-
"All attributes are immutable by default. Under the hood, each model can specify a set of mutable attributes via the `_mutable_attributes` property. In theory this can be overriden. But the general idea is that we will progressively add more and more mutable attributes with time.\n",
335+
"All attributes are immutable by default. Under the hood, each model can specify a set of mutable attributes via the `_mutable_attributes` property. In theory this can be overridden. But the general idea is that we will progressively add more and more mutable attributes with time.\n",
336336
"\n",
337-
"And that concludes this recipe. Arguably, this recipe caters to advanced users, and in particular users who are developping their own models. And yet, one could also argue that modifying parameters of a model on-the-fly is a great tool to have at your disposal when you're doing online machine learning."
337+
"And that concludes this recipe. Arguably, this recipe caters to advanced users, and in particular users who are developing their own models. And yet, one could also argue that modifying parameters of a model on-the-fly is a great tool to have at your disposal when you're doing online machine learning."
338338
]
339339
}
340340
],

docs/recipes/on-hoeffding-trees.ipynb

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@
2626
"\n",
2727
"In this guide, we are going to:\n",
2828
"\n",
29-
"1. summarize the differences accross the multiple HT versions available;\n",
29+
"1. summarize the differences across the multiple HT versions available;\n",
3030
"2. learn how to inspect tree models;\n",
3131
"3. learn how to manage the memory usage of HTs;\n",
3232
"4. compare numerical tree splitters and understand their impact on the iDT induction process.\n",
@@ -888,7 +888,7 @@
888888
"- $n$: Number of observations seen so far.\n",
889889
"- $c$: the number of classes.\n",
890890
"- $s$: the number of split points to evaluate (which means that this is a user-given parameter).\n",
891-
"- $h$: the number of histogram bins or hash slots. Tipically, $h \\ll n$.\n",
891+
"- $h$: the number of histogram bins or hash slots. Typically, $h \\ll n$.\n",
892892
"\n",
893893
"### 4.1. Classification tree splitters\n",
894894
"\n",
@@ -906,7 +906,7 @@
906906
"- The number of split points can be configured in the Gaussian splitter. Increasing this number makes this splitter slower, but it also potentially increases the quality of the obtained query points, implying enhanced tree accuracy. \n",
907907
"- The number of stored bins can be selected in the Histogram splitter. Increasing this number increases the memory footprint and running time of this splitter, but it also potentially makes its split candidates more accurate and positively impacts on the tree's final predictive performance.\n",
908908
"\n",
909-
"Next, we provide a brief comparison of the classification splitters using 10K instances of the Random RBF synthetic dataset. Note that the tree equiped with the Exhaustive splitter does not use Naive Bayes leaves."
909+
"Next, we provide a brief comparison of the classification splitters using 10K instances of the Random RBF synthetic dataset. Note that the tree equipped with the Exhaustive splitter does not use Naive Bayes leaves."
910910
]
911911
},
912912
{

docs/releases/0.12.0.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@
2929
## drift
3030

3131
- Refactor the concept drift detectors to match the remaining of River's API. Warnings are only issued by detectors that support this feature.
32-
- Drifts can be assessed via the property `drift_detected`. Warning signals can be acessed by the property `warning_detected`. The `update` now returns `self`.
32+
- Drifts can be assessed via the property `drift_detected`. Warning signals can be accessed by the property `warning_detected`. The `update` now returns `self`.
3333
- Ensure all detectors automatically reset their inner states after a concept drift detection.
3434
- Streamline `DDM`, `EDDM`, `HDDM_A`, and `HDDM_W`. Make the configurable parameters names match their respective papers.
3535
- Fix bugs in `EDDM` and `HDDM_W`.

docs/releases/0.19.0.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ Calling `learn_one` in a pipeline will now update each part of the pipeline in t
3030
## forest
3131

3232
- Fixed issue with `forest.ARFClassifier` which couldn't be passed a `CrossEntropy` metric.
33-
- Fixed a bug in `forest.AMFClassifier` which slightly improves predictive accurary.
33+
- Fixed a bug in `forest.AMFClassifier` which slightly improves predictive accuracy.
3434
- Added `forest.AMFRegressor`.
3535

3636
## multioutput

docs/releases/0.8.0.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,6 +28,6 @@
2828

2929
## tree
3030

31-
- Unifed base class structure applied to all tree models.
31+
- Unified base class structure applied to all tree models.
3232
- Bug fixes.
3333
- Added `tree.SGTClassifier` and `tree.SGTRegressor`.

0 commit comments

Comments
 (0)