OceanPile: A Large-Scale Multimodal Ocean Corpus for Foundation Models Paper • 2605.00877 • Published 17 days ago • 15
Zero-shot World Models Are Developmentally Efficient Learners Paper • 2604.10333 • Published Apr 11 • 7
Training a Student Expert via Semi-Supervised Foundation Model Distillation Paper • 2604.03841 • Published Apr 4 • 10
Adam's Law: Textual Frequency Law on Large Language Models Paper • 2604.02176 • Published Apr 2 • 501
Rethinking Generalization in Reasoning SFT: A Conditional Analysis on Optimization, Data, and Model Capability Paper • 2604.06628 • Published Apr 8 • 324
DataFlex: A Unified Framework for Data-Centric Dynamic Training of Large Language Models Paper • 2603.26164 • Published Mar 27 • 364
Distilling Human-Aligned Privacy Sensitivity Assessment from Large Language Models Paper • 2603.29497 • Published Mar 31 • 6
CARLA-Air: Fly Drones Inside a CARLA World -- A Unified Infrastructure for Air-Ground Embodied Intelligence Paper • 2603.28032 • Published Mar 30 • 341
Bootstrapping Exploration with Group-Level Natural Language Feedback in Reinforcement Learning Paper • 2603.04597 • Published Mar 4 • 210
From Blind Spots to Gains: Diagnostic-Driven Iterative Training for Large Multimodal Models Paper • 2602.22859 • Published Feb 26 • 151