Rickard Edén

neph1

AI & ML interests

applying, dataset creation, finetuning, loras

Organizations

neph1's activity

posted an update 20 days ago
posted an update 4 months ago
posted an update 5 months ago
view post
Post
592
I've noticed some people are still downloading neph1/sd-seer-griffin-3b
Should I make an update based on a more modern architecture? (griffin-3b is llama (1!))
  • 1 reply
·
posted an update 5 months ago
view post
Post
753
First real version of bellman based on llama 3 instruct 8b has been released!
neph1/llama-3-instruct-bellman-8b-swedish

Close to 16k examples, including 250 rows from my translated codefeedback dataset and a number of non-copyrighted stories.

Two quants are up in the gguf folder. I'll work on adding more quants in the coming days.
reacted to stas's post with 👍 9 months ago
view post
Post
Do you have a hidden massive storage leak thanks to HF hub models and datasets revisions adding up and not getting automatically deleted?

Here is how to delete all old revisions and only keeping main in a few quick steps and no tedious manual editing.

In terminal A:
$ pip install huggingface_hub["cli"] -U
$ huggingface-cli delete-cache --disable-tui
File to edit: /tmp/tmpundr7lky.txt
0 revisions selected counting for 0.0. Continue ? (y/N)

Do not answer the prompt and proceed with my instructions.

(note your tmp file will have a different path, so adjust it below)

In terminal B:
$ cp /tmp/tmpedbz00ox.txt cache.txt
$ perl -pi -e 's|^#(.*detached.*)|$1|' cache.txt
$ cat cache.txt >>  /tmp/tmpundr7lky.txt

The perl one-liner uncommented out all lines that had (detached) in it - so can be wiped out. And then we pasted it back into the tmp file huggingface-cli expects to be edited.

Now go back to terminal A and hit: N, Y, Y, so it looks like:

0 revisions selected counting for 0.0. Continue ? (y/N) n
89 revisions selected counting for 211.7G. Continue ? (y/N) y
89 revisions selected counting for 211.7G. Confirm deletion ? (Y/n) y

Done.

If you messed up with the prompt answering you still have cache.txt file which you can feed again to the new tmp file it'll create when you run huggingface-cli delete-cache --disable-tui again.

For more details and additional techniques please see https://github.com/stas00/ml-engineering/tree/master/storage#huggingface-hub-caches