iandennismiller commited on
Commit
7bde6da
1 Parent(s): 8e4cb30

fix cache path

Browse files
Files changed (1) hide show
  1. bin/llama.sh +22 -13
bin/llama.sh CHANGED
@@ -1,9 +1,10 @@
1
  #!/bin/bash
2
 
3
- # if conf does not exist, create it
4
- if [ ! -f "$HOME/.config/llama/llama-main.conf" ]; then
5
- mkdir -p "$HOME/.config/llama"
6
- cat <<EOF > "$HOME/.config/llama/llama-main.conf"
 
7
  LLAMA_TEMPERATURE=0.1
8
  LLAMA_CONTEXT_SIZE=4096
9
  LLAMA_REPETITION_PENALTY=1.15
@@ -13,13 +14,19 @@ LLAMA_TEMPLATE=chatml
13
  LLAMA_MODEL_NAME=teknium/OpenHermes-2.5-Mistral-7B/openhermes-2.5-mistral-7b-f16.gguf
14
  LLAMA_MODELS_PATH=$HOME/.ai/models/llama
15
  LLAMA_PROMPT_PATH=$HOME/.local/share/llama/prompts
16
- LLAMA_CACHE_PATH=$HOME/.local/share/llama/cache
17
  EOF
18
- fi
19
 
20
- function llama_interactive {
21
  source $HOME/.config/llama/llama-main.conf
22
 
 
 
 
 
 
 
 
23
  # if 4 arguments are provided...
24
  if [[ $# -eq 4 ]]; then
25
  LLAMA_MODEL_NAME=$1
@@ -39,7 +46,7 @@ function llama_interactive {
39
  llama \
40
  --n-gpu-layers 1 \
41
  --model "$LLAMA_MODELS_PATH/$LLAMA_MODEL_NAME" \
42
- --prompt-cache "$LLAMA_CACHE_PATH/$LLAMA_MODEL_NAME" \
43
  --file "$(get_model_prompt $LLAMA_MODEL_NAME)" \
44
  --in-prefix "$(get_model_prefix $LLAMA_TEMPLATE)" \
45
  --in-suffix "$(get_model_suffix $LLAMA_TEMPLATE)" \
@@ -57,7 +64,8 @@ function llama_interactive {
57
  --instruct \
58
  --no-mmap \
59
  --color \
60
- --escape
 
61
  }
62
 
63
  function get_model_prefix {
@@ -69,10 +77,10 @@ function get_model_prefix {
69
  printf "### Instruction: "
70
  ;;
71
  *vicuna*|*wizardlm*|*samantha*)
72
- printf "USER:"
73
  ;;
74
  *based*|*yi*)
75
- printf "Human:"
76
  ;;
77
  *orca*)
78
  printf "### User: "
@@ -95,10 +103,10 @@ function get_model_suffix {
95
  printf "### Response: "
96
  ;;
97
  *vicuna*|*wizardlm*|*samantha*)
98
- printf "ASSISTANT:"
99
  ;;
100
  *based*|*yi*)
101
- printf "Assistant:"
102
  ;;
103
  *orca*)
104
  printf "### Response: "
@@ -135,4 +143,5 @@ function get_model_prompt {
135
  esac
136
  }
137
 
 
138
  llama_interactive $*
 
1
  #!/bin/bash
2
 
3
+ function llama_init_environment {
4
+ # if conf does not exist, create it
5
+ if [ ! -f "$HOME/.config/llama/llama-main.conf" ]; then
6
+ mkdir -p "$HOME/.config/llama"
7
+ cat <<EOF > "$HOME/.config/llama/llama-main.conf"
8
  LLAMA_TEMPERATURE=0.1
9
  LLAMA_CONTEXT_SIZE=4096
10
  LLAMA_REPETITION_PENALTY=1.15
 
14
  LLAMA_MODEL_NAME=teknium/OpenHermes-2.5-Mistral-7B/openhermes-2.5-mistral-7b-f16.gguf
15
  LLAMA_MODELS_PATH=$HOME/.ai/models/llama
16
  LLAMA_PROMPT_PATH=$HOME/.local/share/llama/prompts
17
+ LLAMA_CACHE_PATH=$HOME/.cache/llama
18
  EOF
19
+ fi
20
 
 
21
  source $HOME/.config/llama/llama-main.conf
22
 
23
+ if [ ! -d "$LLAMA_CACHE_PATH" ]; then
24
+ mkdir -p "$LLAMA_CACHE_PATH"
25
+ fi
26
+
27
+ }
28
+
29
+ function llama_interactive {
30
  # if 4 arguments are provided...
31
  if [[ $# -eq 4 ]]; then
32
  LLAMA_MODEL_NAME=$1
 
46
  llama \
47
  --n-gpu-layers 1 \
48
  --model "$LLAMA_MODELS_PATH/$LLAMA_MODEL_NAME" \
49
+ --prompt-cache "$LLAMA_CACHE_PATH/${LLAMA_MODEL_NAME//[\/\.]/-}-${LLAMA_CONTEXT_SIZE}.cache" \
50
  --file "$(get_model_prompt $LLAMA_MODEL_NAME)" \
51
  --in-prefix "$(get_model_prefix $LLAMA_TEMPLATE)" \
52
  --in-suffix "$(get_model_suffix $LLAMA_TEMPLATE)" \
 
64
  --instruct \
65
  --no-mmap \
66
  --color \
67
+ --escape \
68
+ --log-disable
69
  }
70
 
71
  function get_model_prefix {
 
77
  printf "### Instruction: "
78
  ;;
79
  *vicuna*|*wizardlm*|*samantha*)
80
+ printf "USER: "
81
  ;;
82
  *based*|*yi*)
83
+ printf "Human: "
84
  ;;
85
  *orca*)
86
  printf "### User: "
 
103
  printf "### Response: "
104
  ;;
105
  *vicuna*|*wizardlm*|*samantha*)
106
+ printf "ASSISTANT: "
107
  ;;
108
  *based*|*yi*)
109
+ printf "Assistant: "
110
  ;;
111
  *orca*)
112
  printf "### Response: "
 
143
  esac
144
  }
145
 
146
+ llama_init_environment
147
  llama_interactive $*