74 pointsby jinqueeny9 hours ago8 comments
  • kelsey987654317 hours ago
    FYI it also supports pre-training, reward model training and RL, not just fine tuning (sft). My team built a managed solution for training that runs on top of llama factory and it's quite excellent and well supported. You will need pretty serious equipment to get good results out of it, think 8xh200. For people at home i would look at doing an sft of gemma3 270m or maybe a 1.6b qwen3, but keep in mind you have to have the dataset in memory as well as the model and kv-cache. cheers
    • spagettnet6 hours ago
      depends ln your goals of course. but worth mentioning there are plenty of narrowish tasks (think text-to-sql, and other less general language tasks) where llama8b or phi-4 (14b) or even up to 30b with quantization can be trained on 8xa100 with great results. plus these smaller models benefit from being able to be served on a single a100 or even L4 with post training quantization, with wicked fast generation thanks to the lighter model.

      on a related note, at what point are people going to get tired of waiting 20s for an llm to answer their questions? i wish it were more common for smaller models to be used when sufficient.

  • metadat8 hours ago
    This reminds me conceptually of the Nvidia NIM factory where they attempt to optimize models in bulk / en-masse.

    https://www.nvidia.com/en-us/ai/nim-for-manufacturing/

    Word on the street is the project has yielded largely unimpressive results compared to its potential, but NV is still investing in an attempt to further raise the GPU saturation waterline.

    p.s. This project logo stood out to me at presenting the Llama releasing some "steam" with gusto. I wonder if that was intentional? Sorry for the immature take but stopping the scatological jokes is tough.

  • stefanwebb5 hours ago
    There’s a similar library that also includes data synth and LLM-as-a-Judge: https://github.com/oumi-ai/oumi
    • BoorishBears3 hours ago
      Yet another framework lying about Deepseek support.

      I've been trying to actually finetune Deepseek (not distills) and there are few options

  • Twirrim9 hours ago
    https://llamafactory.readthedocs.io/en/latest/

    I found this link more useful.

    "LLaMA Factory is an easy-to-use and efficient platform for training and fine-tuning large language models. With LLaMA Factory, you can fine-tune hundreds of pre-trained models locally without writing any code."

  • sabareesh7 hours ago
    This is great,but most work is involved in curating the dataset and the objective functions for RL.
  • tensorlibb8 hours ago
    This is incredible! What gpu configs, budget to ultra high-end, would you recommend for local fine tuning?

    Always curious to see what other ai enthusiasts are running!

    • spagettnet6 hours ago
      axolotl is great on consumer hardware.
  • jcuenod7 hours ago
    Can you compare this to Unsloth?
  • hall0ween8 hours ago
    are there any use cases, aside from code generation and formatting, where fine-tuning consistently useful?
    • clipclopflop8 hours ago
      Creating small, specialized models for specific tasks. Being able to leverage the up front training/data as a generalized base allows you to quickly create a small local model that can generate outputs for that task that can come close to or match the same you would see in a large/hosted model.