aberrio commited on
Commit
e4263cd
1 Parent(s): 0902b6f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +46 -14
README.md CHANGED
@@ -14,24 +14,56 @@ tags:
14
  - gguf
15
  ---
16
 
17
- ## Model Summary
18
 
19
- The Phi-3-Mini-128K-Instruct is a 3.8 billion-parameter, lightweight, state-of-the-art open model trained using the Phi-3 datasets.
20
- This dataset includes both synthetic data and filtered publicly available website data, with an emphasis on high-quality and reasoning-dense properties.
21
- The model belongs to the Phi-3 family with the Mini version in two variants [4K](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) and [128K](https://huggingface.co/microsoft/Phi-3-mini-128k-instruct) which is the context length (in tokens) that it can support.
22
 
23
- After initial training, the model underwent a post-training process that involved supervised fine-tuning and direct preference optimization to enhance its ability to follow instructions and adhere to safety measures.
24
- When evaluated against benchmarks that test common sense, language understanding, mathematics, coding, long-term context, and logical reasoning, the Phi-3 Mini-128K-Instruct demonstrated robust and state-of-the-art performance among models with fewer than 13 billion parameters.
25
- Resources and Technical Documentation:
 
 
 
 
 
 
 
26
 
27
- + [Phi-3 Microsoft Blog](https://aka.ms/phi3blog-april)
28
- + [Phi-3 Technical Report](https://aka.ms/phi3-tech-report)
29
 
30
- ## Quantized Model Files
 
 
 
31
 
32
- Phi-3 is available in several formats, catering to different computational needs:
 
33
 
34
- - **ggml-model-q8_0.gguf**: 8-bit quantization, providing robust performance with a file size of 3.8 GB.
35
- - **ggml-model-f16.gguf**: Standard 16-bit floating-point format, with a larger file size of 7.2 GB for enhanced precision.
 
 
36
 
37
- These formats, ranging from 4-bit to 16-bit, accommodate various computational environments, from resource-constrained devices to high-end server
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  - gguf
15
  ---
16
 
17
+ ## Phi-3-Mini-128K-Instruct
18
 
19
+ ### Model Information
20
+ Phi-3-Mini-128K-Instruct is a 3.8 billion-parameter, lightweight, instruction-tuned model from Microsoft, belonging to the Phi-3 family. It has been optimized for long-context comprehension and efficient handling of complex, reasoning-dense tasks. The model supports a context length of up to 128K tokens, making it particularly suitable for scenarios involving extended conversations or long-form content generation.
 
21
 
22
+ - **Name**: Phi-3-Mini-128K-Instruct
23
+ - **Parameter Size**: 3.8 billion
24
+ - **Model Family**: Phi-3
25
+ - **Architecture**: Transformer with an enhanced focus on efficient context handling.
26
+ - **Purpose**: Multilingual dialogue generation, text generation, code completion, and summarization.
27
+ - **Training Data**: A combination of synthetic data and filtered, publicly available website data, with an emphasis on reasoning-dense properties.
28
+ - **Supported Languages**: English (primary language).
29
+ - **Release Date**: September 18, 2024
30
+ - **Context Length**: 128K tokens (other versions include a [4K variant](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct))
31
+ - **Knowledge Cutoff**: July 2023
32
 
33
+ ### Quantized Model Files
34
+ Phi-3 is available in several formats, catering to different computational needs and resource constraints:
35
 
36
+ - **ggml-model-q8_0.gguf**: 8-bit quantization, providing robust performance with a file size of 3.8 GB, suitable for resource-constrained environments.
37
+ - **ggml-model-f16.gguf**: 16-bit floating-point format, offering enhanced precision at a larger file size of 7.2 GB.
38
+
39
+ These formats ensure that the Phi-3 Mini-128K can be adapted to a variety of systems, from low-power devices to high-end servers, making it a versatile option for deployments.
40
 
41
+ ### Core Library
42
+ Phi-3-Mini-128K-Instruct can be deployed using `llama.cpp` or `transformers`, with support for high-efficiency long-context inference.
43
 
44
+ - **Primary Framework**: `llama.cpp`
45
+ - **Alternate Frameworks**:
46
+ - `transformers` for integrations into the Hugging Face ecosystem.
47
+ - `vLLM` for efficient inference with optimized memory usage.
48
 
49
+ **Library and Model Links**:
50
+ - **Model Base**: [microsoft/Phi-3-mini-128k-instruct](https://huggingface.co/microsoft/Phi-3-mini-128k-instruct)
51
+ - **Resources and Technical Documentation**:
52
+ - [Phi-3 Microsoft Blog](https://aka.ms/phi3blog-april)
53
+ - [Phi-3 Technical Report](https://aka.ms/phi3-tech-report)
54
+
55
+ ## Safety and Responsible Use
56
+
57
+ The Phi-3-Mini-128K-Instruct is part of the Phi model family, known for its rigorous dataset curation focused on educational and non-toxic sources. Due to its careful design, the Phi-3 series generally avoids generating harmful or biased outputs. This makes it a reliable choice for safety-critical applications and environments where ethical standards are paramount.
58
+
59
+ ### Training Philosophy
60
+ The Phi-3 series models are intentionally trained on textbooks, research papers, and high-quality language corpora, avoiding sources that might introduce harmful, biased, or inappropriate content. As a result, Phi-3 maintains a strong adherence to safe and controlled responses, even when handling sensitive topics or instructions.
61
+
62
+ ### Risk Profile and Use Recommendations
63
+ While no AI model is entirely risk-free, Phi-3's safety features minimize the likelihood of producing unwanted or offensive outputs. However, it is still recommended that users conduct scenario-specific testing to verify its behavior in deployment environments. For additional confidence, consider the following guidelines:
64
+
65
+ - **Intended Use**: Education, research, and general-purpose dialogue systems.
66
+ - **Deployment**: Suitable for low-risk applications where adherence to ethical and safety guidelines is crucial.
67
+ - **Community Testing and Feedback**: Open to user feedback to improve safety benchmarks further and align with best practices.
68
+
69
+ For more information on Phi's safety approach, refer to [Phi-3 Technical Report](https://aka.ms/phi3-tech-report).