R6410418/Jackrong-llm-finetuning-guide
Projects and tools related to fine-tuning pre-trained AI models.
LlamaIndex is the leading document agent and OCR platform
Unsloth Studio is a web UI for training and running open models like Qwen3.5, Gemma 4, DeepSeek, gpt-oss locally.
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
Multi-lingual large voice generation model, providing inference, training and deployment full-stack ability.
从 NLP 到 LLM 的算法全栈教程,在线阅读地址:https://datawhalechina.github.io/base-llm/
Developer Asset Hub for NVIDIA Nemotron — A one-stop resource for training recipes, usage cookbooks, datasets, and full end-to-end reference examples to build with Nemotron models
Your Cheat Sheet for AI Engineering Interview – Questions and Answers.
Zero-friction LLM fine-tuning skill for Claude Code, Gemini CLI & any ACP agent. Unsloth on NVIDIA · TRL+MPS/MLX on Apple Silicon. Automates env setup, LoRA training (SFT, DPO, GRPO, vision), post-hoc GRPO log diagnostics, evaluation, and export end-to-end. Part of the Gaslamp AI platform.
System Level Intelligent Router for Mixture-of-Models at Cloud, Data Center and Edge
🦞 Just talk to your agent — it learns and EVOLVES 🧬.
A simple, performant and scalable Jax LLM!
Fast ML inference & training for ONNX models in Rust
Scalable data pre processing and curation toolkit for LLMs
One-stop handbook for building, deploying, and understanding LLM agents with 60+ skeletons, tutorials, ecosystem guides, and evaluation tools.
VeritasGraph: Enterprise-Grade Graph RAG for Secure, On-Premise AI with Verifiable Attribution
Full-Stack Development Platform for Building Reliable Agents
🏭 AI agent platform with skills for protein engineering, the noob-friendly AI tutorial tool for life science professionals.
Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)
A framework for agentic tool use training with reinforcement learning
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
The Unity Machine Learning Agents Toolkit (ML-Agents) is an open-source project that enables games and simulations to serve as environments for training intelligent agents using deep reinforcement learning and imitation learning.
Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We also show you how to solve end to end problems using Llama model family and using them on various provider services
Machine Learning Engineering Open Book
Transfer learning / domain adaptation / domain generalization / multi-task learning etc. Papers, codes, datasets, applications, tutorials.-迁移学习
A powerful tool for creating datasets for LLM fine-tuning 、RAG and Eval
1 Line of code data quality profiling & exploratory data analysis for Pandas and Spark DataFrames.
Run any open-source LLMs, such as DeepSeek and Llama, as OpenAI compatible API endpoint in the cloud.
Go ahead and axolotl questions
The AI developer platform. Use Weights & Biases to train and fine-tune models, and manage models from experimentation to production.
Example 📓 Jupyter notebooks that demonstrate how to build, train, and deploy machine learning models using 🧠 Amazon SageMaker.
Easily fine-tune, evaluate and deploy gpt-oss, Qwen3, DeepSeek-R1, or any open source LLM / VLM!
Using Low-rank adaptation to quickly fine-tune diffusion models.
Dynamic, resilient AI orchestration. Coordinate data, models, and compute as you build AI workflows. Flyte 2 now available locally: https://github.com/flyteorg/flyte-sdk
Build, Evaluate, and Optimize AI Systems. Includes evals, RAG, agents, fine-tuning, synthetic data generation, dataset management, MCP, and more.
《大语言模型》作者:赵鑫,李军毅,周昆,唐天一,文继荣
RAG (Retrieval Augmented Generation) Framework for building modular, open source applications for production by TrueFoundry
Transfer Learning Library for Domain Adaptation, Task Adaptation, and Domain Generalization
Open Source Pre-training Model Framework in PyTorch & Pre-trained Model Zoo
OpenDeepWiki is the open-source version of the DeepWiki project, aiming to provide a powerful knowledge management and collaboration platform. The project is mainly developed using C# and TypeScript, supporting modular design, and is easy to expand and customize.
OneTrainer is a one-stop solution for all your Diffusion training needs.
streamline the fine-tuning process for multimodal models: PaliGemma 2, Florence-2, and Qwen2.5-VL
Learn to build your Second Brain AI assistant with LLMs, agents, RAG, fine-tuning, LLMOps and AI systems techniques.
Mastering Applied AI, One Concept at a Time
Control plane for agents and engineers to provision compute and run training and inference across NVIDIA, AMD, TPU, and Tenstorrent GPUs—on clouds, Kubernetes, and bare-metal clusters.
Distributed AI Model Training and LLM Fine-Tuning on Kubernetes
Custom Diffusion: Multi-Concept Customization of Text-to-Image Diffusion (CVPR 2023)
A JAX research toolkit for building, editing, and visualizing neural networks.
jiant is an nlp toolkit
Synthetic data curation for post-training and structured data extraction
Ultrafast serverless GPU inference, sandboxes, and background jobs
Code for TKDE paper "Self-supervised learning on graphs: Contrastive, generative, or predictive"
Hypernetworks that adapt LLMs for specific benchmark tasks using only textual task description as the input
Generate High-Quality Synthetics, Train, Measure, and Evaluate in a Single Pipeline
Official repository of my book "A Hands-On Guide to Fine-Tuning LLMs with PyTorch and Hugging Face"
AI molecular design tool for de novo design, scaffold hopping, R-group replacement, linker design and molecule optimization.
Model Merging in LLMs, MLLMs, and Beyond: Methods, Theories, Applications and Opportunities. ACM Computing Surveys, 2026.
Personal Project: MPP-Qwen14B & MPP-Qwen-Next(Multimodal Pipeline Parallel based on Qwen-LM). Support [video/image/multi-image] {sft/conversations}. Don't let the poverty limit your imagination! Train your own 8B/14B LLaVA-training-like MLLM on RTX3090/4090 24GB.
Complete-Life-Cycle-of-a-Data-Science-Project
This repository provides programs to build Retrieval Augmented Generation (RAG) code for Generative AI with LlamaIndex, Deep Lake, and Pinecone leveraging the power of OpenAI and Hugging Face models for generation and evaluation.
🏗️ Fine-tune, build, and deploy open-source LLMs easily!
Build computer vision models in a fraction of the time and with less data.
Collection of awesome parameter-efficient fine-tuning resources.
A WebGL accelerated JavaScript library for training and deploying ML models.
Distributed training framework for TensorFlow, Keras, PyTorch, and Apache MXNet.
Collection of scripts to aggregate image data for the purposes of training an NSFW Image Classifier
Low-code framework for building custom LLMs, neural networks, and other AI models
H2O LLM Studio - a framework and no-code GUI for fine-tuning LLMs. Documentation: https://docs.h2o.ai/h2o-llmstudio/
Open Source Data Security Platform for Developers to Monitor and Detect PII, Anonymize Production Data and Sync it across environments.
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
Fine-tuning ChatGLM-6B with PEFT | 基于 PEFT 的高效 ChatGLM 微调
🔥🔥High-Performance Face Recognition Library on PaddlePaddle & PyTorch🔥🔥
AI Agent that handles engineering tasks end-to-end: integrates with developers’ tools, plans, executes, and iterates until it achieves a successful result.
A library for transfer learning by reusing parts of TensorFlow models.
OML 1.0 via Fingerprinting: Open, Monetizable, and Loyal AI
The next generation deep reinforcement learning tookit
🦖 𝗟𝗲𝗮𝗿𝗻 about 𝗟𝗟𝗠𝘀, 𝗟𝗟𝗠𝗢𝗽𝘀, and 𝘃𝗲𝗰𝘁𝗼𝗿 𝗗𝗕𝘀 for free by designing, training, and deploying a real-time financial advisor LLM system ~ 𝘴𝘰𝘶𝘳𝘤𝘦 𝘤𝘰𝘥𝘦 + 𝘷𝘪𝘥𝘦𝘰 & 𝘳𝘦𝘢𝘥𝘪𝘯𝘨 𝘮𝘢𝘵𝘦𝘳𝘪𝘢𝘭𝘴
Collection of AWESOME vision-language models for vision tasks
LLM Finetuning with peft
A general fine-tuning kit geared toward image/video/audio diffusion models.
Build, personalize and control your own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJEk6
Deep learning PyTorch library for time series forecasting, classification, and anomaly detection (originally for flood forecasting).
Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models
Your Automatic Prompt Engineering Assistant for GenAI Applications
FedML - The Research and Production Integrated Federated Learning Library: https://fedml.ai
A repository that contains models, datasets, and fine-tuning techniques for DB-GPT, with the purpose of enhancing model performance in Text-to-SQL
A comprehensive guide to building RAG-based LLM applications for production.
[ICLR 2025] LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs
🦄 State-of-the-Art Conversational AI with Transfer Learning
:dart: Task-oriented embedding tuning for BERT, CLIP, etc.
Adapting Meta AI's Segment Anything to Downstream Tasks with Adapters and Prompts
The official implementation of Self-Play Fine-Tuning (SPIN)
Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models"
DataDreamer: Prompt. Generate Synthetic Data. Train & Align Models. 🤖💤
Tencent Pre-training framework in PyTorch & Pre-trained Model Zoo
[TPAMI 2023] LibFewShot: A Comprehensive Library for Few-shot Learning.
ICCV 2023-2025 Papers: Discover cutting-edge research from ICCV 2023-25, the leading computer vision conference. Stay updated on the latest in computer vision and deep learning, with code included. ⭐ support visual intelligence development!
Pre-training of Deep Bidirectional Transformers for Language Understanding: pre-train TextCNN
A complete guide to start and improve your LLM skills in 2026 with little background in the field and stay up-to-date with the latest news and state-of-the-art techniques!