shafire commited on
Commit
36eb076
1 Parent(s): 9f6f292

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +78 -3
README.md CHANGED
@@ -1,3 +1,78 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ tags:
4
+ - autotrain
5
+ - text-generation-inference
6
+ - text-generation
7
+ - peft
8
+ library_name: transformers
9
+ base_model: meta-llama/Meta-Llama-3.1-8B
10
+ widget:
11
+ - messages:
12
+ - role: user
13
+ content: What is your favorite condiment?
14
+ license: apache-2.0
15
+ ---
16
+
17
+ talktoaiQ - SkynetZero LLM
18
+
19
+ talktoaiQ aka SkynetZero is a quantum-powered language model trained with reflection datasets and custom TalkToAI datasets. The model went through several iterations, including re-writing of datasets and validation phases, due to errors encountered during testing and conversion into a fully functional LLM. This iterative process ensures SkynetZero can handle complex, multi-dimensional reasoning tasks with an emphasis on ethical decision-making.
20
+
21
+ Key Highlights of talktoaiQ:
22
+ - Advanced Quantum Reasoning: Integration of quantum-inspired math systems enables talktoaiQ to tackle complex ethical dilemmas and multi-dimensional problem-solving tasks.
23
+ - Custom Re-Written Datasets: The training involved multiple rounds of AI-assisted dataset curation, where reflection datasets were re-written for clarity, accuracy, and consistency. Additionally, TalkToAI datasets were integrated and re-processed to align with talktoaiQ’s quantum reasoning framework.
24
+ - Iterative Improvement: During testing and model conversion, the datasets were re-written and validated several times to address errors. Each iteration enhanced the model’s ethical consistency and problem-solving accuracy.
25
+ - Fine-Tuned on LLaMA 3.1 8B: The model was fine-tuned on the LLaMA 3.1 8B architecture, integrating multiple specialized datasets to ensure high-quality text generation capabilities.
26
+
27
+ Model Overview
28
+ - Developed by: Shafaet Brady Hussain - researchforum.online
29
+ - Funded by: Researchforum.online
30
+ - Shared by: TalkToAI - https://talktoai.org
31
+ - Language(s): English
32
+ - Model type: Causal Language Model
33
+ - Fine-tuned from: LLaMA 3.1 8B (Meta)
34
+ - License: Apache-2.0
35
+
36
+ Usage:
37
+ You can use the following code snippet to load and interact with talktoaiQ:
38
+
39
+ from transformers import AutoModelForCausalLM, AutoTokenizer
40
+
41
+ model_path = "PATH_TO_THIS_REPO"
42
+
43
+ tokenizer = AutoTokenizer.from_pretrained(model_path)
44
+ model = AutoModelForCausalLM.from_pretrained(
45
+ model_path,
46
+ device_map="auto",
47
+ torch_dtype="auto"
48
+ ).eval()
49
+
50
+ # Prompt content: "hi"
51
+ messages = [
52
+ {"role": "user", "content": "hi"}
53
+ ]
54
+
55
+ input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
56
+ output_ids = model.generate(input_ids.to("cuda"))
57
+ response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True)
58
+
59
+ # Model response: "Hello! How can I assist you today?"
60
+ print(response)
61
+
62
+ Training Methodology
63
+ talktoaiQ was fine-tuned on the LLaMA 3.1 8B architecture using custom datasets. The datasets underwent AI-assisted re-writing to enhance clarity and consistency. Throughout the training process, emphasis was placed on multi-variable quantum reasoning and ensuring alignment with ethical decision-making principles. After identifying errors during testing and conversion, datasets were further improved across multiple epochs.
64
+
65
+ - Training Regime: Mixed Precision (fp16)
66
+ - Training Duration: 8 hours on a high-performance GPU server
67
+
68
+ Further Research and Contributions
69
+ talktoaiQ is part of an ongoing effort to explore AI-human co-creation in the development of quantum-enhanced AI models. Collaboration with OpenAI’s Agent Zero played a significant role in curating, editing, and validating datasets, pushing the boundaries of what large language models can achieve.
70
+
71
+ - Contributions: https://researchforum.online
72
+ - Contact: [email protected]
73
+
74
+ Carbon Emissions & Environmental Impact:
75
+ - Hardware Used: AMD EPYC CPU and High-End GPU
76
+ - Hours Used: 8 hours
77
+ - Compute Region: On-premise
78
+ - Carbon Emissions: Estimated 20 kg CO2