forked from dusty-nv/jetson-containers
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathconfig.py
34 lines (24 loc) · 866 Bytes
/
config.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
import copy
# ggml version tracks fork
ggml = copy.deepcopy(package)
ggml['name'] = 'llama_cpp:ggml'
ggml['build_args'] = {
'LLAMA_CPP_PYTHON_REPO': 'dusty-nv/llama-cpp-python',
'LLAMA_CPP_PYTHON_BRANCH': 'v0.1.78a',
}
ggml['test'].extend([
"test_model.py --model $(huggingface-downloader TheBloke/Llama-2-7B-GGML/llama-2-7b.ggmlv3.q4_0.bin)",
"test_tokenizer.py --model $(huggingface-downloader TheBloke/Llama-2-7B-GGML/llama-2-7b.ggmlv3.q4_0.bin)"
])
# gguf version tracks main
gguf = copy.deepcopy(package)
gguf['name'] = 'llama_cpp:gguf'
gguf['alias'] = 'llama_cpp'
gguf['build_args'] = {
'LLAMA_CPP_PYTHON_REPO': 'abetlen/llama-cpp-python',
'LLAMA_CPP_PYTHON_BRANCH': 'main',
}
gguf['test'].extend([
"test_model.py --model $(huggingface-downloader TheBloke/Llama-2-7B-GGUF/llama-2-7b.Q4_K_S.gguf)"
])
package = [ggml, gguf]