Skip to content

Commit b2f8c1a

Browse files
authored
Merge pull request #19 from StochasticTree/typo-hotfix
Fix typo in tree overview
2 parents 92cf76d + 189eff7 commit b2f8c1a

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

docs/about.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,7 @@ f(X_i) = g_1(X_i) + \dots + g_k(X_i)
5959

6060
There are several ways to train an ensemble of decision trees (sometimes called "forests"), the most popular of which are [random forests](https://en.wikipedia.org/wiki/Random_forest) and
6161
[gradient boosting](https://en.wikipedia.org/wiki/Gradient_boosting). Their main difference is that random forests train
62-
all $m$ trees independently of one another, while boosting trains tree sequentially, so that tree $j$ depends on the result of training trees 1 through $j-1$.
62+
all $m$ trees independently of one another, while boosting trains trees sequentially, so that tree $j$ depends on the result of training trees 1 through $j-1$.
6363
Libraries like [xgboost](https://xgboost.readthedocs.io/en/stable/) and [LightGBM](https://lightgbm.readthedocs.io/en/latest/) are popular examples of boosted tree ensembles.
6464

6565
Tree ensembles often [outperform neural networks and other machine learning methods on tabular datasets](https://arxiv.org/abs/2207.08815),

0 commit comments

Comments
 (0)