|
--- |
|
inference: false |
|
license: other |
|
language: |
|
- en |
|
model_name: Samantha 1.1 Llama 33B |
|
quantized_by: iandennismiller |
|
base_model: ehartford/samantha-1.1-llama-33b |
|
datasets: |
|
- ehartford/samantha-data |
|
--- |
|
|
|
# Eric Hartford's Samantha 1.1 Llama 33B GGUF |
|
|
|
Samantha 1.1 is a very smart model. |
|
For those of us with 24gb usable RAM (including Apple Silicon with 32gb RAM), this is just about the best model available as of October, 2023. |
|
|
|
I've converted the model to GGUF for the sake of compatibility. |
|
Currently, only Q4_K_S is available because that's the largest model that runs in 24gb. |
|
|
|
For more information, see [Eric Hartford's Samantha 1.1 Llama 33B](https://huggingface.co/ehartford/samantha-1.1-llama-33b). |
|
To examine the original quant, see [TheBloke/samantha-1.1-llama-33B-GGML](https://huggingface.co/TheBloke/samantha-1.1-llama-33B-GGML) |
|
|
|
The particular quants selected for this repo are in support of [calm](https://github.com/iandennismiller/calm), which is a language model runner that automatically uses the right prompts, templates, context size, etc. |
|
|