Skip to content

PaperCodex

Subscribe

Instruction Tuning

Self-Instruct: Bootstrap High-Quality Instruction Data Without Human Annotations

Self-Instruct: Bootstrap High-Quality Instruction Data Without Human Annotations 4557

For teams building or fine-tuning large language models (LLMs), one of the biggest bottlenecks is the scarcity of high-quality, diverse…

01/05/2026Instruction Tuning, Synthetic Data Generation, Zero-shot Generalization
FlipVQA-Miner: Automatically Extract High-Quality Visual QA Pairs from Textbooks for Reliable LLM Training

FlipVQA-Miner: Automatically Extract High-Quality Visual QA Pairs from Textbooks for Reliable LLM Training 1737

Large Language Models (LLMs) and multimodal systems increasingly demand high-quality, human-authored supervision data—especially for tasks requiring reasoning, visual understanding, and…

01/04/2026Educational Data Mining, Instruction Tuning, Visual Question Answering
LLaMA-Adapter: Efficiently Transform LLaMA into Instruction-Following or Multimodal AI with Just 1.2M Parameters

LLaMA-Adapter: Efficiently Transform LLaMA into Instruction-Following or Multimodal AI with Just 1.2M Parameters 5907

If you’re working on a project that requires a capable language model—but lack the GPU budget, time, or infrastructure for…

12/26/2025Instruction Tuning, Multimodal Learning, Parameter-Efficient Fine-Tuning
UltraChat: Train Powerful Open-Source Chat Models with 1.5M High-Quality, Privacy-Safe AI Dialogues

UltraChat: Train Powerful Open-Source Chat Models with 1.5M High-Quality, Privacy-Safe AI Dialogues 2721

If you’re a technical decision-maker evaluating options for building or fine-tuning a conversational AI system, you know that high-quality instruction-following…

12/26/2025Conversational AI, Instruction Tuning, Multi-turn Dialogue Modeling
WizardCoder: Open-Source Code LLM That Outperforms ChatGPT and Gemini in Code Generation

WizardCoder: Open-Source Code LLM That Outperforms ChatGPT and Gemini in Code Generation 9472

WizardCoder is a state-of-the-art open-source Code Large Language Model (Code LLM) that delivers exceptional performance on code generation tasks—often surpassing…

12/26/2025Code Generation, Instruction Tuning, Programming Assistance
Instruction Pre-Training: Boost Language Model Performance from Day One with Supervised Multitask Pre-Training

Instruction Pre-Training: Boost Language Model Performance from Day One with Supervised Multitask Pre-Training 4150

Traditional language model (LM) development follows a two-stage process: unsupervised pre-training on massive raw text corpora, followed by instruction tuning…

12/19/2025Instruction Tuning, Language Model Pre-training, Multitask Learning
Align Anything: The First Open Framework for Aligning Any-to-Any Multimodal Models with Human Intent

Align Anything: The First Open Framework for Aligning Any-to-Any Multimodal Models with Human Intent 4562

As AI systems grow more capable across diverse data types—text, images, audio, and video—the challenge of aligning them with human…

12/19/2025Instruction Tuning, Multimodal Alignment, Reinforcement Learning From Human Feedback
Copyright © 2026 PaperCodex.
  • Facebook
  • YouTube
  • Twitter

PaperCodex