user: | danielhanchen |
created: | Sep 8, 2021 |
karma: | 843 |
about: | Unsloth github.com/unslothai/unsloth - finetune Llama 2x faster + use 70% less VRAM 1. Used to work at NVIDIA RAPIDS cuML 2. Discord: https://discord.gg/unsloth 3. Github: https://github.com/danielhanchen 4. Twitter / X: x.com/danielhanchen 5. Email: my handle @ gmail.com 6. Bug fixes for Gemma: https://news.ycombinator.com/item?id=39671146 7. Bug fixes for Gradient Accumulation: https://x.com/danielhanchen/status/1846235913443262891?lang=en |