Hacker News
new
top
best
ask
show
job
1.8-3.3x faster Embedding finetuning now in Unsloth
(
unsloth.ai
)
3 points
by
electroglyph
16 days ago
3 comments
storystarling
15 days ago
Do the memory savings carry over to inference or is this strictly optimizing the backward pass? I'm running embedding pipelines via Celery and being able to squeeze this into lower VRAM would help the margins quite a bit.
danielhanchen
16 days ago
Excited to have collabed on this! Thanks electroglyph for the contrib!
electroglyph
16 days ago
see also:
https://www.reddit.com/r/LocalLLaMA/comments/1qk9vmv/1833x_f...