Highlights
- Pro
Stars
AI-Native Studio for everything audio. Sound Design/Audio Editing/AI Producer/Fast Exports. Built for web, runs anywhere.
[NeurIPS 2025 D&B Spotlight] Scaling Data for SWE-agents
DeepEP: an efficient expert-parallel communication library
Ring attention implementation with flash attention
Organize the Web: Constructing Domains Enhances Pre-Training Data Curation
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
A PyTorch native platform for training generative AI models
Fully open reproduction of DeepSeek-R1
ripgrep recursively searches directories for a regex pattern while respecting your gitignore
Aioli: A unified optimization framework for language model data mixing
A bibliography and survey of the papers surrounding o1
16-fold memory access reduction with nearly no loss
SGLang is a high-performance serving framework for large language models and multimodal models.
Official repository for DistFlashAttn: Distributed Memory-efficient Attention for Long-context LLMs Training
Cache efficient implementation for Latent Dirichlet Allocation
Convert PDF to markdown + JSON quickly with high accuracy
Official repository for MATES: Model-Aware Data Selection for Efficient Pretraining with Data Influence Models [NeurIPS 2024]
[ACL 2024 (Oral)] A Prospector of Long-Dependency Data for Large Language Models
Emerge is a browser-based interactive codebase and dependency visualization tool for many different programming languages. It supports some basic code quality and graph metrics and provides a simpl…
SWE-agent takes a GitHub issue and tries to automatically fix it, using your LM of choice. It can also be employed for offensive cybersecurity or competitive coding challenges. [NeurIPS 2024]
[ICML 2024] Selecting High-Quality Data for Training Language Models
The official implementation of Self-Play Fine-Tuning (SPIN)





