YokaiKoibito's picture
Update README.md
50cba23
metadata
license: llama2
datasets:
  - ehartford/wizard_vicuna_70k_unfiltered
tags:
  - uncensored
  - wizard
  - vicuna
  - llama

This is an fp16 copy of jarradh/llama2_70b_chat_uncensored for faster downloading and less disk space usage than the fp32 original. I simply imported the model to CPU with torch_dtype=torch.float16 and then exported it again. I also added a chat_template entry derived from the model card to the tokenizer_config.json file, which previously didn't have one. All credit for the model goes to jarradh.

Arguable a better name for this model would be something like Llama-2-70B_Wizard-Vicuna-Uncensored-fp16, but to avoid confusion I'm sticking with jarradh's naming scheme.

Repositories available

Prompt template: Human-Response

### HUMAN:
{prompt}

### RESPONSE: