Diving deep into the world of Large Language Models! From architecture exploration to deployment, I'm passionate about understanding and building with these fascinating models. Proud owner of an RTX 4090 that powers my local LLM experiments! π
- Exploring various LLM architectures
- Fine-tuning techniques (LoRA, QLoRA, PEFT)
- Building with agentic frameworks (LangGraph, AutoGPT)
- Optimizing inference for consumer GPUs
- Core: Python, PyTorch, Transformers
- LLM Tools:
- π€ Hugging Face ecosystem
- π¦ LangChain, LlamaIndex
- π vLLM, text-generation-inference
- Quantization: bitsandbytes, GPTQ, AWQ
- Deployment: FastAPI, Gradio, Docker
- Hardware: NVIDIA RTX 4090 (precious β¨)
- Advanced techniques in Prompt Engineering
- Multi-agent systems and autonomous AI
- Knowledge Graphs for RAG (Retrieval-Augmented Generation) optimization
- Model distillation and compression methods
- Vector databases (QDrant, Pinecone)
TBA