New ask Hacker News story: Ask HN: What data are you using to finetune LLMs?
Ask HN: What data are you using to finetune LLMs?
2 by rahimnathwani | 1 comments on Hacker News.
There have been a number of posts on HN recently about how to fine tune LLMs. These posts talk mainly about: - different methods of finetuning (full retraining, LoRA) - different base models - data sets (e.g. Alpaca) - objectives (creativity, instruction following) I haven't seen much discussion about people finetuning an LLM on domain-specific data, e.g. - medical records - standup comedy jokes - internal corporate data So, are any of you fine-tuning your LLMs using such niche data? I'd love to hear about your experiences and motivations! Even if you're working with proprietary datasets, I'm still interested. After all, knowing what you're doing won't allow us to duplicate it, as we don't have access to the same data.
2 by rahimnathwani | 1 comments on Hacker News.
There have been a number of posts on HN recently about how to fine tune LLMs. These posts talk mainly about: - different methods of finetuning (full retraining, LoRA) - different base models - data sets (e.g. Alpaca) - objectives (creativity, instruction following) I haven't seen much discussion about people finetuning an LLM on domain-specific data, e.g. - medical records - standup comedy jokes - internal corporate data So, are any of you fine-tuning your LLMs using such niche data? I'd love to hear about your experiences and motivations! Even if you're working with proprietary datasets, I'm still interested. After all, knowing what you're doing won't allow us to duplicate it, as we don't have access to the same data.
Comments
Post a Comment