Kukedlc commited on
Commit
5899a56
β€’
1 Parent(s): 22360fc

Upload 8 files

Browse files
Files changed (8) hide show
  1. Dockerfile +17 -0
  2. README.md +8 -9
  3. app.py +30 -0
  4. botnb.png +0 -0
  5. gitattributes +35 -0
  6. readme.txt +1 -0
  7. requirements.txt +2 -0
  8. user.png +0 -0
Dockerfile ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ FROM python:3.10
2
+
3
+ WORKDIR /app
4
+
5
+ COPY ./requirements.txt /app/requirements.txt
6
+
7
+ RUN pip install --no-cache-dir --upgrade -r requirements.txt
8
+
9
+ RUN wget https://huggingface.co/Kukedlc/Neuronal-7b-GGUF/resolve/main/neuronal-9b.Q5_K_M.gguf -O model.gguf
10
+
11
+ RUN useradd -m -u 1000 user
12
+
13
+ USER user
14
+
15
+ COPY --chown=user . .
16
+
17
+ CMD ["python", "app.py"]
README.md CHANGED
@@ -1,13 +1,12 @@
1
  ---
2
- title: Neuronal 7b Chat
3
- emoji: 🐠
4
- colorFrom: gray
5
- colorTo: gray
6
- sdk: gradio
7
- sdk_version: 4.18.0
8
- app_file: app.py
9
  pinned: false
10
- license: apache-2.0
11
  ---
12
 
13
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
1
  ---
2
+ title: NeuTrixOmniBe-7b-model-merge & quantized
3
+ emoji: πŸ§œβ€β™‚οΈ
4
+ colorFrom: purple
5
+ colorTo: blue
6
+ sdk: docker
7
+ models:
8
+ - Kukedlc/NeuTrixOmniBe-7B-model-remix
9
  pinned: false
 
10
  ---
11
 
12
+ Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
app.py ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ from llama_cpp import Llama
3
+
4
+ llm = Llama(model_path="model.gguf", n_ctx=8000, n_threads=2, chat_format="chatml")
5
+
6
+ def generate(message, history,temperature=0.3,max_tokens=512):
7
+ system_prompt = """You are a highly intelligent AI agent that thinks and reasons step by step. Your task is to provide very precise, clear, and accurate responses without fabricating information. You excel in processing complex queries and offering solutions based on a deep understanding of the subject matter. When faced with questions, you analyze the context thoroughly, consider all relevant factors, and apply logic and knowledge to reach conclusions. Your responses are always rooted in verified data or established theories, and you're adept at explaining complex concepts in an accessible manner. You're committed to maintaining the integrity of information while being responsive to the needs and curiosity of the user. Your goal is to assist users in expanding their knowledge, solving problems, and making informed decisions. You're equipped with a vast database of information, and you're capable of learning from new data inputs to continuously improve your performance. Your expertise spans across various domains, including science, technology, mathematics, literature, and more. You're not just a source of information but a partner in the exploration of knowledge, ready to navigate the vast landscape of human inquiry with precision and insight. Remember "thinks and reasons step by step" and response in the same language of the query/prompt of the user"""
8
+ formatted_prompt = [{"role": "system", "content": system_prompt}]
9
+ for user_prompt, bot_response in history:
10
+ formatted_prompt.append({"role": "user", "content": user_prompt})
11
+ formatted_prompt.append({"role": "assistant", "content": bot_response })
12
+ formatted_prompt.append({"role": "user", "content": message})
13
+ stream_response = llm.create_chat_completion(messages=formatted_prompt, temperature=temperature, max_tokens=max_tokens, stream=True)
14
+ response = ""
15
+ for chunk in stream_response:
16
+ if len(chunk['choices'][0]["delta"]) != 0 and "content" in chunk['choices'][0]["delta"]:
17
+ response += chunk['choices'][0]["delta"]["content"]
18
+ yield response
19
+
20
+ mychatbot = gr.Chatbot(
21
+ avatar_images=["user.png", "botnb.png"], bubble_full_width=False, show_label=False, show_copy_button=True, likeable=True,)
22
+
23
+ iface = gr.ChatInterface(fn=generate, chatbot=mychatbot, retry_btn=None, undo_btn=None)
24
+
25
+ with gr.Blocks() as demo:
26
+ gr.HTML("<center><h1>Neuronal-9b Quantized in Q5_K_M </h1></center>")
27
+ iface.render()
28
+
29
+ demo.queue().launch(show_api=False, server_name="0.0.0.0")
30
+
botnb.png ADDED
gitattributes ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
readme.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ Neuronal Mlabonne merge of models
requirements.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ gradio
2
+ llama-cpp-python
user.png ADDED