Coding-LLM are trained on old data. Even the latest GPT-4 Turbo Code Interpreter (CI) has a knowledge cut-off at April 2023. All AI research from the last 7 moths are not in the training data of commercial coding LLMs. And RAG lines of code do not help at all, given the complex interdependencies of code libs.
Therefore an elegant solution for AI researcher is to fine-tune your own Coding-LLM on the latest GitHub repos and coding data. Which is exactly the content of this video: How to fine-tune your personal coding-LLM (or a Co-pilot like Microsoft's GitHub co-pilot or any CODE-LLM like StarCoder).
We code a classical fine-tuning of a Code LLM (StarCoder), then code a PEFT-LoRA tuning for a personal Code LLM and also code a QLoRA tuning based on a special code dataset and a pre-trained Code LLM.
We compare compute infrastructure requirements for distributed computing (8x A100 80GB GPU) and apply DeepSpeed and fully shared Data Parallel (FSDP) for memory optimization.
Python files available for download at (all rights with authors of GitHub repos /HuggingFace wizards)):
-----------------------------------------------------------------------------------------------------------------------------------------
TRAIN.PY
https://github.com/pacman100/DHS-LLM-...
DEEPSPEED
https://github.com/pacman100/DHS-LLM-...
FSDP
https://github.com/pacman100/DHS-LLM-...
PEFT LoRA
https://github.com/pacman100/DHS-LLM-...
Complete Colab NB for PEFT LoRA Quantization Fine-tuning my personal CODE LLM:
https://colab.research.google.com/dri...
#ai
#coding
#pythonprogramming