Skip to content

PaperCodex

Subscribe

Language Modeling

JetMoE: High-Performance LLMs Under $100K—Open, Efficient, and Accessible

JetMoE: High-Performance LLMs Under $100K—Open, Efficient, and Accessible 985

Building powerful language models used to be the exclusive domain of well-funded tech giants. But JetMoE is changing that narrative.…

01/13/2026Efficient Inference, Language Modeling, Sparse Mixture-of-experts
SpargeAttention: Universal, Training-Free Sparse Attention for Faster LLM, Image & Video Inference Without Retraining

SpargeAttention: Universal, Training-Free Sparse Attention for Faster LLM, Image & Video Inference Without Retraining 814

Large AI models—from language generators to video diffusion systems—are bottlenecked by the attention mechanism, whose computational cost scales quadratically with…

01/13/2026Image Generation, Language Modeling, Video Generation
LLaMA-MoE: High-Performance Mixture-of-Experts LLM with Only 3.5B Active Parameters

LLaMA-MoE: High-Performance Mixture-of-Experts LLM with Only 3.5B Active Parameters 994

If you’re a developer, researcher, or technical decision-maker working with large language models (LLMs), you’ve likely faced a tough trade-off:…

01/13/2026Efficient Inference, Language Modeling, Text Generation
Copyright © 2026 PaperCodex.
  • Facebook
  • YouTube
  • Twitter

PaperCodex