Skip to content

Commit 64cb5c3

Browse files
Merge pull request #181 from Shohail-Ismail/patch-1
Add on timing advantage under Summary; No code changes, therefore aurtomerge tests will run only after commit.
2 parents 6dfb427 + 3f4e3d9 commit 64cb5c3

File tree

1 file changed

+21
-3
lines changed

1 file changed

+21
-3
lines changed

README.md

Lines changed: 21 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ The Cerebros package is an ultra-precise Neural Architecture Search (NAS) / Auto
66

77
## Cerebros Community Edition and Cerebros Enterprise
88

9-
The Cerebros community edition provides an open-source minimum viable single parameter set NAS and also also provides an example manifest for an exhaustive Neural Architecture Search to run on Kubeflow/Katib. This is licensed for free use provided that the use is consistent with the ethical use provisions in the license described at the bottom of this page. You can easily reproduce this with the Jupyter notebook in the directory `/kubeflow-pipeline`, using the Kale Jupyter notebook extension. For a robust managed neural architecture search experience hosted on Google Cloud Platform and supported by our SLA, we recommend Cerebros Enterprise, our commercial version. Soon you will be able to sign up and immediately start using it at `https://www.cerebros.one`. In the meantime, we can set up your own Cerbros managed neural architecture search pipeline for you with a one business day turnaround. We offer consulting, demos, full service machine learning service and can provision you with your own full neural architecture search pipeline complete with automated Bayesian hyperparameter search. Contact David Thrower:`[email protected]` or call us at (US country code 1) `(650) 789-4375`. Additionally, we can complete machine learning tasks for your organization. Give us a call.
9+
The Cerebros community edition provides an open-source minimum viable single parameter set NAS and also provides an example manifest for an exhaustive Neural Architecture Search to run on Kubeflow/Katib. This is licensed for free use provided that the use is consistent with the ethical use provisions in the license described at the bottom of this page. You can easily reproduce this with the Jupyter notebook in the directory `/kubeflow-pipeline`, using the Kale Jupyter notebook extension. For a robust managed neural architecture search experience hosted on Google Cloud Platform and supported by our SLA, we recommend Cerebros Enterprise, our commercial version. Soon you will be able to sign up and immediately start using it at `https://www.cerebros.one`. In the meantime, we can set up your own Cerebros managed neural architecture search pipeline for you with a one business day turnaround. We offer consulting, demos, full service machine learning service and can provision you with your own full neural architecture search pipeline complete with automated Bayesian hyperparameter search. Contact David Thrower:`[email protected]` or call us at (US country code 1) `(650) 789-4375`. Additionally, we can complete machine learning tasks for your organization. Give us a call.
1010

1111

1212

@@ -26,7 +26,7 @@ In a biological brain, neurons connect in a multi-dimensional lattice of vertica
2626

2727
That is what we did here. We built a neural architecture search that connects Dense layers in this manner.
2828

29-
What if we made a multi-layer pereceptron that looks like this: (Green triangles are Keras Input layers. Blue Squares are Keras Concatenate layers. The Pink stretched ovals are Keras Dense layers. The one stretched red oval is the network's Output layer. It is presumed that there is a batch normaliation layer between each Concatenate layer and the Dense layer it feeds into.)
29+
What if we made a multi-layer perceptron that looks like this: (Green triangles are Keras Input layers. Blue Squares are Keras Concatenate layers. The Pink stretched ovals are Keras Dense layers. The one stretched red oval is the network's Output layer. It is presumed that there is a batch normalisation layer between each Concatenate layer and the Dense layer it feeds into.)
3030

3131
![assets/Brain-lookalike1.png](assets/Brain-lookalike1.png)
3232

@@ -77,7 +77,7 @@ Best model name: 2023_01_12_23_42_cerebros_auto_ml_test_meta_0/models/tr_0000000
7777
...
7878
7979
```
80-
80+
---
8181
## Summary of Results
8282

8383
- Ames housing data set, not pre-processed or scaled, non-numerical columns dropped:
@@ -88,6 +88,24 @@ Best model name: 2023_01_12_23_42_cerebros_auto_ml_test_meta_0/models/tr_0000000
8888

8989
For further details, see ![documentation/examples/use-example-detailed.md](documentation/examples/use-example-detailed.md)
9090

91+
### Practical O(n) timing with increasing sequence length
92+
93+
Recent updates replaced the text embedding base model with an interleaved Rotary Positional Embedding (iRoPE) in the text-classification proof of concept. This change allows the model to handle longer sequences without the quadratic slow-down common to many transformer architectures.
94+
95+
Benchmarks show that training time grows in proportion to sequence length, while validation accuracy stays stable:
96+
97+
| seq_len | val_binary_accuracy | min/model | total_min | timing_relative_to_1024 | commit |
98+
|---:|:---:|---:|---:|---:|:--|
99+
| 3072 | 0.955 | 65.942 | 329.715 | 2.817 | [`4bc217b`](https://github.com/david-thrower/cerebros-core-algorithm-alpha/commit/4bc217b36d1baf8b9a81fe53000e9e13e6e87801) |
100+
| 1536 | 0.960 | 37.270 | 186.360 | 1.591 | [`286ba81`](https://github.com/david-thrower/cerebros-core-algorithm-alpha/commit/286ba81a1e51493d748ded727bd602a4398248a8) |
101+
| 1024 | 0.952 | 23.420 | 117.080 | 1.000 | [`9893bfc`](https://github.com/david-thrower/cerebros-core-algorithm-alpha/commit/9893bfc55d4f7b753eff79e0c5c70e4992c61085) |
102+
103+
104+
The `timing_relative_to_1024` column is calculated as `min/model(seq_len) / min/model(1024)`.
105+
For examale, 1024 to 3072 tokens is roughly x3 in sequence length and x2.82 in time, which is close to linear scaling once fixed overhead is considered.
106+
107+
This outcome follows earlier work on more scalable tokenisation, RoPE/iRoPE integration, and related performance fixes.
108+
91109
# Documentation
92110

93111
![documentation/api-docs/summary-and-navigation.md](documentation/api-docs/summary-and-navigation.md)

0 commit comments

Comments
 (0)