Text Generation
Transformers
PyTorch
llama
text-generation-inference
Inference Endpoints

Set model_max_length of tokenizer equal to the max_length of the model

#26
by zxcvvxcz - opened
Ready to merge
This branch is ready to get merged automatically.

Sign up or log in to comment