--- license: apache-2.0 language: - en metrics: - accuracy library_name: transformers pipeline_tag: text-generation datasets: - Barishni-blinchik/uwbruh tags: - kawaii - cringe --- ***Some cringe...* Oh well, hello!** I present GPT2 to you, but with a bit of kawaii. Chat template ``` <|USER|> Hello <|ASSISTANT|> ``` --- # Training Results The following metrics are from the latest training session of our model: ## Overview - **Global Step:** 615 - **Training Loss:** 0.1303 ## Detailed Metrics - **Training Runtime:** 413.1481 seconds - **Training Samples per Second:** 5.947 - **Training Steps per Second:** 1.489 - **Total Floating Point Operations (FLOs):** 641,994,522,624,000.0 - **Training Loss:** 0.13032278840134784 - **Epoch:** 3.0 ---