Update ChatApp/app.py
Browse files- ChatApp/app.py +15 -15
ChatApp/app.py
CHANGED
@@ -33,35 +33,35 @@ top_directory = os.path.dirname(os.path.dirname(os.path.realpath(__file__)))
|
|
33 |
tokenizer_path = os.path.join(top_directory, "tokenizer.model")
|
34 |
|
35 |
available_models = {
|
36 |
-
"Llama-2 7B Float16": {
|
37 |
"onnx_file": os.path.join(
|
38 |
-
top_directory, "FP16", "
|
39 |
),
|
40 |
"tokenizer_path": tokenizer_path,
|
41 |
-
"embedding_file": os.path.join(top_directory, "embeddings.pth"),
|
42 |
},
|
43 |
-
"Llama-2 7B
|
44 |
"onnx_file": os.path.join(
|
45 |
-
top_directory, "FP32", "
|
46 |
),
|
47 |
"tokenizer_path": tokenizer_path,
|
48 |
-
"embedding_file": os.path.join(
|
49 |
-
|
50 |
-
|
51 |
-
},
|
52 |
-
"Llama-2 7B Chat Float16": {
|
53 |
"onnx_file": os.path.join(
|
54 |
-
top_directory, "FP16
|
55 |
),
|
56 |
"tokenizer_path": tokenizer_path,
|
57 |
-
"embedding_file": os.path.join(top_directory, "embeddings
|
58 |
},
|
59 |
-
"Llama-2 7B
|
60 |
"onnx_file": os.path.join(
|
61 |
-
top_directory, "FP32
|
62 |
),
|
63 |
"tokenizer_path": tokenizer_path,
|
64 |
-
"embedding_file": os.path.join(
|
|
|
|
|
65 |
},
|
66 |
}
|
67 |
|
|
|
33 |
tokenizer_path = os.path.join(top_directory, "tokenizer.model")
|
34 |
|
35 |
available_models = {
|
36 |
+
"Llama-2 7B Chat Float16": {
|
37 |
"onnx_file": os.path.join(
|
38 |
+
top_directory, "FP16-Chat", "LlamaV2_7B_FT_float16.onnx"
|
39 |
),
|
40 |
"tokenizer_path": tokenizer_path,
|
41 |
+
"embedding_file": os.path.join(top_directory, "embeddings-chat.pth"),
|
42 |
},
|
43 |
+
"Llama-2 7B Chat Float32": {
|
44 |
"onnx_file": os.path.join(
|
45 |
+
top_directory, "FP32-Chat", "LlamaV2_7B_FT_float32.onnx"
|
46 |
),
|
47 |
"tokenizer_path": tokenizer_path,
|
48 |
+
"embedding_file": os.path.join(top_directory, "embeddings-chat.pth"),
|
49 |
+
},
|
50 |
+
"Llama-2 7B Float16": {
|
|
|
|
|
51 |
"onnx_file": os.path.join(
|
52 |
+
top_directory, "FP16", "LlamaV2_7B_float16.onnx"
|
53 |
),
|
54 |
"tokenizer_path": tokenizer_path,
|
55 |
+
"embedding_file": os.path.join(top_directory, "embeddings.pth"),
|
56 |
},
|
57 |
+
"Llama-2 7B Float32": {
|
58 |
"onnx_file": os.path.join(
|
59 |
+
top_directory, "FP32", "LlamaV2_7B_float32.onnx"
|
60 |
),
|
61 |
"tokenizer_path": tokenizer_path,
|
62 |
+
"embedding_file": os.path.join(
|
63 |
+
top_directory, "embeddings.pth"
|
64 |
+
),
|
65 |
},
|
66 |
}
|
67 |
|