Update README.md
Browse files
README.md
CHANGED
@@ -14,24 +14,56 @@ tags:
|
|
14 |
- gguf
|
15 |
---
|
16 |
|
17 |
-
##
|
18 |
|
19 |
-
|
20 |
-
|
21 |
-
The model belongs to the Phi-3 family with the Mini version in two variants [4K](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) and [128K](https://huggingface.co/microsoft/Phi-3-mini-128k-instruct) which is the context length (in tokens) that it can support.
|
22 |
|
23 |
-
|
24 |
-
|
25 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
26 |
|
27 |
-
|
28 |
-
|
29 |
|
30 |
-
|
|
|
|
|
|
|
31 |
|
32 |
-
|
|
|
33 |
|
34 |
-
- **
|
35 |
-
- **
|
|
|
|
|
36 |
|
37 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
- gguf
|
15 |
---
|
16 |
|
17 |
+
## Phi-3-Mini-128K-Instruct
|
18 |
|
19 |
+
### Model Information
|
20 |
+
Phi-3-Mini-128K-Instruct is a 3.8 billion-parameter, lightweight, instruction-tuned model from Microsoft, belonging to the Phi-3 family. It has been optimized for long-context comprehension and efficient handling of complex, reasoning-dense tasks. The model supports a context length of up to 128K tokens, making it particularly suitable for scenarios involving extended conversations or long-form content generation.
|
|
|
21 |
|
22 |
+
- **Name**: Phi-3-Mini-128K-Instruct
|
23 |
+
- **Parameter Size**: 3.8 billion
|
24 |
+
- **Model Family**: Phi-3
|
25 |
+
- **Architecture**: Transformer with an enhanced focus on efficient context handling.
|
26 |
+
- **Purpose**: Multilingual dialogue generation, text generation, code completion, and summarization.
|
27 |
+
- **Training Data**: A combination of synthetic data and filtered, publicly available website data, with an emphasis on reasoning-dense properties.
|
28 |
+
- **Supported Languages**: English (primary language).
|
29 |
+
- **Release Date**: September 18, 2024
|
30 |
+
- **Context Length**: 128K tokens (other versions include a [4K variant](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct))
|
31 |
+
- **Knowledge Cutoff**: July 2023
|
32 |
|
33 |
+
### Quantized Model Files
|
34 |
+
Phi-3 is available in several formats, catering to different computational needs and resource constraints:
|
35 |
|
36 |
+
- **ggml-model-q8_0.gguf**: 8-bit quantization, providing robust performance with a file size of 3.8 GB, suitable for resource-constrained environments.
|
37 |
+
- **ggml-model-f16.gguf**: 16-bit floating-point format, offering enhanced precision at a larger file size of 7.2 GB.
|
38 |
+
|
39 |
+
These formats ensure that the Phi-3 Mini-128K can be adapted to a variety of systems, from low-power devices to high-end servers, making it a versatile option for deployments.
|
40 |
|
41 |
+
### Core Library
|
42 |
+
Phi-3-Mini-128K-Instruct can be deployed using `llama.cpp` or `transformers`, with support for high-efficiency long-context inference.
|
43 |
|
44 |
+
- **Primary Framework**: `llama.cpp`
|
45 |
+
- **Alternate Frameworks**:
|
46 |
+
- `transformers` for integrations into the Hugging Face ecosystem.
|
47 |
+
- `vLLM` for efficient inference with optimized memory usage.
|
48 |
|
49 |
+
**Library and Model Links**:
|
50 |
+
- **Model Base**: [microsoft/Phi-3-mini-128k-instruct](https://huggingface.co/microsoft/Phi-3-mini-128k-instruct)
|
51 |
+
- **Resources and Technical Documentation**:
|
52 |
+
- [Phi-3 Microsoft Blog](https://aka.ms/phi3blog-april)
|
53 |
+
- [Phi-3 Technical Report](https://aka.ms/phi3-tech-report)
|
54 |
+
|
55 |
+
## Safety and Responsible Use
|
56 |
+
|
57 |
+
The Phi-3-Mini-128K-Instruct is part of the Phi model family, known for its rigorous dataset curation focused on educational and non-toxic sources. Due to its careful design, the Phi-3 series generally avoids generating harmful or biased outputs. This makes it a reliable choice for safety-critical applications and environments where ethical standards are paramount.
|
58 |
+
|
59 |
+
### Training Philosophy
|
60 |
+
The Phi-3 series models are intentionally trained on textbooks, research papers, and high-quality language corpora, avoiding sources that might introduce harmful, biased, or inappropriate content. As a result, Phi-3 maintains a strong adherence to safe and controlled responses, even when handling sensitive topics or instructions.
|
61 |
+
|
62 |
+
### Risk Profile and Use Recommendations
|
63 |
+
While no AI model is entirely risk-free, Phi-3's safety features minimize the likelihood of producing unwanted or offensive outputs. However, it is still recommended that users conduct scenario-specific testing to verify its behavior in deployment environments. For additional confidence, consider the following guidelines:
|
64 |
+
|
65 |
+
- **Intended Use**: Education, research, and general-purpose dialogue systems.
|
66 |
+
- **Deployment**: Suitable for low-risk applications where adherence to ethical and safety guidelines is crucial.
|
67 |
+
- **Community Testing and Feedback**: Open to user feedback to improve safety benchmarks further and align with best practices.
|
68 |
+
|
69 |
+
For more information on Phi's safety approach, refer to [Phi-3 Technical Report](https://aka.ms/phi3-tech-report).
|