Jared Van Bortel
commited on
Commit
•
f414603
1
Parent(s):
613d71b
update README
Browse files
README.md
CHANGED
@@ -15,7 +15,7 @@ tags:
|
|
15 |
---
|
16 |
|
17 |
***
|
18 |
-
**
|
19 |
***
|
20 |
|
21 |
<br/>
|
@@ -31,7 +31,7 @@ This repo contains llama.cpp-compatible files for [nomic-embed-text-v1.5](https:
|
|
31 |
|
32 |
llama.cpp will default to 2048 tokens of context with these files. To use the full 8192 tokens that Nomic Embed is benchmarked on, you will have to choose a context extension method. The original model uses Dynamic NTK-Aware RoPE scaling, but that is not currently available in llama.cpp. A combination of YaRN and linear scaling is an acceptable substitute.
|
33 |
|
34 |
-
These files were converted and quantized with llama.cpp commit [
|
35 |
|
36 |
## Example `llama.cpp` Command
|
37 |
|
|
|
15 |
---
|
16 |
|
17 |
***
|
18 |
+
**Note**: For compatiblity with current llama.cpp, please download the files published on 2/15/2024. The files originally published here do not work after llama.cpp [PR 5500](https://github.com/ggerganov/llama.cpp/pull/5500).
|
19 |
***
|
20 |
|
21 |
<br/>
|
|
|
31 |
|
32 |
llama.cpp will default to 2048 tokens of context with these files. To use the full 8192 tokens that Nomic Embed is benchmarked on, you will have to choose a context extension method. The original model uses Dynamic NTK-Aware RoPE scaling, but that is not currently available in llama.cpp. A combination of YaRN and linear scaling is an acceptable substitute.
|
33 |
|
34 |
+
These files were converted and quantized with llama.cpp [PR 5500](https://github.com/ggerganov/llama.cpp/pull/5500), commit [34aa045de](https://github.com/ggerganov/llama.cpp/pull/5500/commits/34aa045de44271ff7ad42858c75739303b8dc6eb).
|
35 |
|
36 |
## Example `llama.cpp` Command
|
37 |
|