HACKER Q&A
📣 Eric_BB

Guides, books or repos for LLM fine-tuning


Hi everyone, I’m a student working on my thesis. I need to do classification using LLM fine-tuning and have access to my university’s HPC resources. I’m looking for guides, books or repositories that can help me understand how to do LLM fine-tuning. I don’t have a budget for training using APIs, but I’m interested in understanding what the cost would look like. Does anyone have any suggestions? Thank you so much!


  👤 extasia Accepted Answer ✓
I'd recommend huggingface's transformers library. You probably don't wanna be finetuning these LLMs yourself since that's a big endeavour, but secondly because they already have the NLU capabilities to solve your problem. You basically want to train a single classification layer on top of your models output. I believe it'll be something like "CausalLMForClassification" in the huggingface docs.

As far as resources go for finetuning I haven't found many great resources. But again, you probably don't wanna be finetuning these massive LLMs for your particular use case -\_/- Buenos suertes.


👤 ilaksh
Please state the assignment in detail. What are you classifying exactly, for example.

If it's just some simple text that is fairly regular then you might not need actual fine tuning. You could just use the OpenAI API.

https://github.com/leehanchung/lora-instruct

What exactly are the HPC resources. Are they GPUs and what type and how many.