Edit model card

mlx-community/Maxine-34B-stock

This model was converted to MLX format from louisbrulenaudet/Maxine-34B-stock using mlx-vlm version 0.15.2. Refer to the original model card for more details on the model.

Use with mlx

pip install -U mlx-vlm
python -m mlx_vlm.generate --model  mlx-community/Maxine-34B-stock --max-tokens 100 --temp 0.0
from mlx_lm import load, generate

model, tokenizer = load("mlx-community/Maxine-34B-stock")
response = generate(model, tokenizer, prompt="hello", verbose=True)

Citing & Authors

If you use this code in your research, please use the following BibTeX entry.

@misc{louisbrulenaudet2024,
  author =       {Louis Brulé Naudet},
  title =        {Maxine-34B-stock, an xtraordinary 34B model},
  year =         {2024}
  howpublished = {\url{https://huggingface.co/mlx-community/Maxine-34B-stock}},
}

Feedback

If you have any feedback, please reach out at [email protected].

Downloads last month
3
Safetensors
Model size
5.76B params
Tensor type
FP16
·
U32
·
Inference Examples
Inference API (serverless) is not available, repository is disabled.

Model tree for mlx-community/Maxine-34B-stock

Finetuned
this model

Datasets used to train mlx-community/Maxine-34B-stock