Skip to content

PaperCodex

Subscribe
Self-Instruct: Bootstrap High-Quality Instruction Data Without Human Annotations

Self-Instruct: Bootstrap High-Quality Instruction Data Without Human Annotations 4557

For teams building or fine-tuning large language models (LLMs), one of the biggest bottlenecks is the scarcity of high-quality, diverse…

01/05/2026Instruction Tuning, Synthetic Data Generation, Zero-shot Generalization
TTRL: Boost LLM Reasoning Without Labels Using Test-Time Reinforcement Learning

TTRL: Boost LLM Reasoning Without Labels Using Test-Time Reinforcement Learning 836

Imagine being able to improve a large language model’s (LLM) reasoning capabilities after deployment, using only unlabeled test data—no ground-truth…

01/05/2026Reasoning, Reinforcement Learning, Test-time Scaling
RecAI: Prevent Out-of-Domain Recommendations with Plug-and-Play LLM Integration for Accurate, Explainable, and Interactive Systems

RecAI: Prevent Out-of-Domain Recommendations with Plug-and-Play LLM Integration for Accurate, Explainable, and Interactive Systems 913

Large Language Models (LLMs) have unlocked new possibilities for recommender systems—enabling natural-language interactions, personalized explanations, and dynamic user control. Yet…

01/05/2026Constrained Generation, Explainable AI, Recommender Systems
FinTeam: A Multi-Agent Financial Intelligence System That Generates Human-Accepted Reports and Outperforms GPT-4o

FinTeam: A Multi-Agent Financial Intelligence System That Generates Human-Accepted Reports and Outperforms GPT-4o 779

Financial analysis is rarely a solo endeavor. In real-world institutions—from investment banks to asset management firms—complex tasks like producing quarterly…

01/05/2026Financial Reasoning, Multi-agent Systems, Retrieval-Augmented Generation
MAGI-1: Autoregressive Video Generation at Scale with Constant Memory and Real-Time Streaming

MAGI-1: Autoregressive Video Generation at Scale with Constant Memory and Real-Time Streaming 530

MAGI-1 is a breakthrough world model designed for autoregressive video generation at scale. Unlike conventional video diffusion or transformer-based approaches…

01/05/2026Autoregressive Modeling, Image-to-video, Video Generation
HiDream-I1: Generate and Edit High-Quality Images in Seconds with Sparse Diffusion Transformer

HiDream-I1: Generate and Edit High-Quality Images in Seconds with Sparse Diffusion Transformer 777

The rapid evolution of AI-driven image generation has unlocked incredible creative potential—but often at a steep cost: slow inference, massive…

01/05/2026Instruction-based Image Editing, Multimodal Generative Modeling, Text-to-Image Generation
ZipVoice-Dialog: Generate Realistic Spoken Dialogues Instantly—No Fine-Tuning, No Templates

ZipVoice-Dialog: Generate Realistic Spoken Dialogues Instantly—No Fine-Tuning, No Templates 662

Creating natural-sounding spoken dialogues between two people has long been a pain point in AI-driven voice applications. Traditional approaches either…

01/05/2026Non-autoregressive TTS, Spoken Dialogue Generation, Zero-shot Text-to-Speech
YOLOv13: Boost Real-Time Object Detection Accuracy Without Sacrificing Speed or Efficiency

YOLOv13: Boost Real-Time Object Detection Accuracy Without Sacrificing Speed or Efficiency 827

For engineers, researchers, and product teams building real-time vision systems—whether for surveillance cameras, autonomous drones, or mobile apps—achieving high detection…

01/05/2026Edge AI, Object Detection, Real-time Computer Vision
UniAnimate-DiT: High-Fidelity Human Animation from a Single Image and Pose Sequence – No Full Retraining Needed

UniAnimate-DiT: High-Fidelity Human Animation from a Single Image and Pose Sequence – No Full Retraining Needed 797

Animating a static human image into a realistic, temporally coherent video used to require massive datasets, complex pipelines, or retraining…

01/05/2026Diffusion Transformer, Human Image Animation, Video Generation
360-LLaMA-Factory: Plug-and-Play Sequence Parallelism for Long-Context SFT and DPO Without Rewriting Your Workflow

360-LLaMA-Factory: Plug-and-Play Sequence Parallelism for Long-Context SFT and DPO Without Rewriting Your Workflow 571

Training large language models (LLMs) on long sequences—whether for document-level instruction tuning, multi-modal reasoning, or complex alignment tasks—has long been…

01/05/2026Direct Preference Optimization, Long-Context Training, Supervised Fine-tuning

Posts pagination

Previous 1 … 5 6 7 … 43 Next
Copyright © 2026 PaperCodex.
  • Facebook
  • YouTube
  • Twitter

PaperCodex