prince-canuma
commited on
Commit
•
fe11376
1
Parent(s):
9d82e24
Update README.md
Browse files
README.md
CHANGED
@@ -270,8 +270,11 @@ Fig 5. Performance comparision of Llama-3-8B, Llama-2-13B, Yi-1.5-6B and Llama-3
|
|
270 |
|
271 |
Llama-3-6B is competive with model within it's category and upto 2x larger than it self across 6 diverse benchmarks.
|
272 |
|
273 |
-
#### Summary
|
274 |
|
|
|
|
|
|
|
275 |
|
276 |
|
277 |
## Citation
|
@@ -288,6 +291,17 @@ Llama-3-6B is competive with model within it's category and upto 2x larger than
|
|
288 |
}
|
289 |
```
|
290 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
291 |
|
292 |
## References:
|
293 |
|
|
|
270 |
|
271 |
Llama-3-6B is competive with model within it's category and upto 2x larger than it self across 6 diverse benchmarks.
|
272 |
|
273 |
+
#### Summary and future directions:
|
274 |
|
275 |
+
This experiment was a success! Using this technique, I'll be able to build many variants. This is the first of many new base models I intend to create.
|
276 |
+
|
277 |
+
Next, I plan to explore different data mixtures and perform full fine-tuning, all of which will contribute to developing other small model as well as larger and more robust models.
|
278 |
|
279 |
|
280 |
## Citation
|
|
|
291 |
}
|
292 |
```
|
293 |
|
294 |
+
# **Thank You!**
|
295 |
+
|
296 |
+
I want to extend my heartfelt thanks to the community for their invaluable expertise and unwavering support.
|
297 |
+
|
298 |
+
This is my most ambitious project yet, and it wouldn't have been possible without the incredible open-source ML community!
|
299 |
+
|
300 |
+
Developers, I am eager to see and hear about the innovative fine-tunes and applications you create.
|
301 |
+
|
302 |
+
Users, I am excited to learn about your experiences and use cases.
|
303 |
+
|
304 |
+
Thank you for your interest and support!
|
305 |
|
306 |
## References:
|
307 |
|