Fine-Tuning vs RAG – How to Choose the Right Approach to Training LLMs on Your Data Post date July 29, 2025 Post author By Manish Shivanandhan Post categories In fine-tuning-llms, fine-tuning-vs-rag, llm-fine-tuning, llms, rag, rag-vs-fine-tuning, training-llms-on-your-data, what-is-rag
Batched Prompting for Efficient GPT-4 Annotatio Post date April 18, 2025 Post author By Language Models (dot tech) Post categories In ai-feedback-loops, ai-preference-optimization, contrastive-learning-ai, direct-nash-optimization, dno-algorithm, how-to-train-ai, llm-fine-tuning, rhlf-optimization
Understanding Concentrability in Direct Nash Optimization Post date April 17, 2025 Post author By Language Models (dot tech) Post categories In ai-feedback-loops, ai-preference-optimization, contrastive-learning-ai, direct-nash-optimization, dno-algorithm, how-to-train-ai, llm-fine-tuning, rhlf-optimization
Extending Direct Nash Optimization for Regularized Preferences Post date April 17, 2025 Post author By Language Models (dot tech) Post categories In ai-feedback-loops, ai-preference-optimization, contrastive-learning-ai, direct-nash-optimization, dno-algorithm, how-to-train-ai, llm-fine-tuning, rhlf-optimization
What Does the Future of AI Model Training Hold? Post date April 17, 2025 Post author By Language Models (dot tech) Post categories In ai-feedback-loops, ai-preference-optimization, contrastive-learning-ai, direct-nash-optimization, dno-algorithm, how-to-train-ai, llm-fine-tuning, rhlf-optimization
Exploring Cutting-Edge Approaches to Iterative LLM Fine Tuning Post date April 16, 2025 Post author By Language Models (dot tech) Post categories In ai-feedback-loops, ai-preference-optimization, contrastive-learning-ai, direct-nash-optimization, dno-algorithm, how-to-train-ai, llm-fine-tuning, rhlf-optimization
AI That Trains Itself? Here’s How it Works Post date April 16, 2025 Post author By Language Models (dot tech) Post categories In ai-feedback-loops, ai-preference-optimization, contrastive-learning-ai, direct-nash-optimization, dno-algorithm, how-to-train-ai, llm-fine-tuning, rhlf-optimization
Direct Nash Optimization Beats Bigger Models with Better Data Post date April 15, 2025 Post author By Language Models (dot tech) Post categories In ai-feedback-loops, ai-preference-optimization, contrastive-learning-ai, direct-nash-optimization, dno-algorithm, how-to-train-ai, llm-fine-tuning, rhlf-optimization
The Art of Arguing With Yourself—And Why It’s Making AI Smarter Post date April 15, 2025 Post author By Language Models (dot tech) Post categories In ai-feedback-loops, ai-preference-optimization, contrastive-learning-ai, direct-nash-optimization, hackernoon-top-story, how-to-train-ai, llm-fine-tuning, rhlf-optimization
Fine-Tuning LLaMA for Multi-Stage Text Retrieval: Conclusion, Acknowledgements and References Post date July 5, 2024 Post author By Writings, Papers and Blogs on Text Models Post categories In bi-encoder-architecture, fine-tuning-llama, llama, llm-fine-tuning, multi-stage-text-retrieval, rankllama, repllama, transformer-architecture
Related Work on Fine-Tuning LLaMA for Multi-Stage Text Retrieval Post date July 5, 2024 Post author By Writings, Papers and Blogs on Text Models Post categories In bi-encoder-architecture, fine-tuning-llama, llama, llm-fine-tuning, multi-stage-text-retrieval, rankllama, repllama, transformer-architecture
Fine-Tuning LLaMA for Multi-Stage Text Retrieval: Experiments Post date July 5, 2024 Post author By Writings, Papers and Blogs on Text Models Post categories In bi-encoder-architecture, fine-tuning-llama, llama, llm-fine-tuning, multi-stage-text-retrieval, rankllama, repllama, transformer-architecture
Optimizing Text Retrieval Pipelines with LLaMA Models Post date July 5, 2024 Post author By Writings, Papers and Blogs on Text Models Post categories In bi-encoder-architecture, fine-tuning-llama, llama, llm-fine-tuning, multi-stage-text-retrieval, rankllama, repllama, transformer-architecture
Fine-Tuning LLaMA for Multi-Stage Text Retrieval Post date July 5, 2024 Post author By Writings, Papers and Blogs on Text Models Post categories In bi-encoder-architecture, fine-tuning-llama, hackernoon-top-story, llama, llm-fine-tuning, multi-stage-text-retrieval, rankllama, transformer-architecture
YaFSDP – An LLM Training Tool That Cuts GPU Usage by 20% – Is Out Now Post date June 22, 2024 Post author By Yandex Post categories In good-company, gpu-utilization, imporve-llm-training, llm-fine-tuning, llm-optimization, llm-training, open-source-tools, what-is-yafsdp