Where is the mixtral-8x7b's tokenizer encoder? Is there a specific repository or node module?
1
#180 opened 8 months ago
by
RamanSB
What is the max token limit on this model?
2
#179 opened 8 months ago
by
RamanSB
Finetuning Mixtral 8x7B Instruct-v0.1 using Transformers
2
#178 opened 8 months ago
by
Ateeqq
Update chat template to resemble the prompt as stated in the model card.
7
#176 opened 8 months ago
by
nilsec
max_sequence_length
1
#175 opened 8 months ago
by
Ravnoor1
Awesome. I Got Very Good Responses, However...
#174 opened 8 months ago
by
deleted
How to run the full model ?
2
#171 opened 8 months ago
by
dounykim
Is there a working/quantized/exl2 (etc) version that will fit on a single 24GB video card (4090)
2
#170 opened 8 months ago
by
cleverest
403 error
1
#169 opened 8 months ago
by
minhphan-qbe
Adding Evaluation Results
#168 opened 8 months ago
by
leaderboard-pr-bot
Rename README.md to RegulusOne
#167 opened 8 months ago
by
Theguy666
Help: CUDA Out of Memory. Hardware requirements.
2
#147 opened 9 months ago
by
zebfreeman
Update README.md
#146 opened 9 months ago
by
frank76rm
Experimental use
#144 opened 9 months ago
by
yassineelkhadiri14
TemplateError: Conversation roles must alternate user/assistant/user/assistant/...
4
#143 opened 9 months ago
by
quamer23
Is instruction format necessary
2
#142 opened 9 months ago
by
supercharge19
[AUTOMATED] Model Memory Requirements
3
#141 opened 9 months ago
by
model-sizer-bot
Update README.md
#140 opened 9 months ago
by
woodyk
Cuda Out of memory issue when deploying mistralai/Mixtral-8x7B-Instruct-v0.1 on AWS "ml.g5.48xlarge"
1
#139 opened 9 months ago
by
sonalisbapte
slow response
1
#138 opened 9 months ago
by
bhavanam2809
Sparsity in mixtral
#137 opened 9 months ago
by
dpk17
Request: DOI
#136 opened 9 months ago
by
Sonny03
Running in Multi-gpu's
5
#134 opened 9 months ago
by
kmukeshreddy
Update README.md
#133 opened 9 months ago
by
gmverbas
How to format custom dataset to finetune Mixtral with TRL SFT script?
#132 opened 9 months ago
by
icpro
How to use run the code on Colab Free Tier or Mac OS?
16
#131 opened 9 months ago
by
dounykim
Different answer after each request
2
#130 opened 9 months ago
by
amin2557
How to finetune the model?
2
#129 opened 9 months ago
by
akasranjan
How much Resource is needed to run the Mixtral ?
1
#128 opened 9 months ago
by
rkhapre
Update README.md
#126 opened 9 months ago
by
mariakatosvich
The inference API Endpoint gives wrongly formatted answer based on the context given but works well in example Spaces. How we can fix this?
9
#125 opened 9 months ago
by
rkhapre
Request: DOI
#124 opened 9 months ago
by
jsr2
Update README.md
#123 opened 9 months ago
by
Pawamami
what is max input token limit of this model?
1
#122 opened 9 months ago
by
vaidehirao
Request: SDFSDFSD
1
#120 opened 9 months ago
by
seedeera
Consistency check failed - model-00019-of-00019.safetensors
#118 opened 9 months ago
by
br1-pist
Difference in EOS token between Mistral/Mixtral and LLAMA.
1
#117 opened 9 months ago
by
xkszltl
Model Output is Changed
9
#116 opened 9 months ago
by
AnzaniAI
The chat template doesn't support a system prompt
6
#114 opened 9 months ago
by
sam-kap
How to get 'output_router_logits'
1
#113 opened 10 months ago
by
cts13
Run inference on 2 GPUs
1
#112 opened 10 months ago
by
bweinstein123
Running a 4-bit Quantized 7B Model on a PC: Feasibility and Insights
4
#109 opened 10 months ago
by
edw-hug-face
How to get result as good as Hugging Face Chat Mixtral-8x7b-Instruct
9
#107 opened 10 months ago
by
Panckackes
FR Discutions ouverte
3
#105 opened 10 months ago
by
Pablito2fois
Rapport d'étonnement
#104 opened 10 months ago
by
YannCHANET
How to use transfimer
#100 opened 10 months ago
by
sethdwumah