All tags
Topic: "representation-learning"
Yann LeCun’s AMI Labs launches with a $1.03B seed to build world models around JEPA
Yann LeCun launched Advanced Machine Intelligence (AMI Labs) with a record $1.03B seed round at a $3.5B pre-money valuation, aiming to build AI models that understand the physical world through world models rather than just language prediction. The startup, based in Europe with locations in Paris and Zürich, is framed as a major milestone for European AI and backed by a prominent founding team including Alex Lebrun, Saining Xie, and Pascale Fung. The mission is described as a "long-term scientific endeavor" to create AI that "perceives, learns, reasons and acts" in the real world.
not much happened today
qwen3-vl-4b qwen3-vl-8b qwen2.5-vl-72b deepseek-v3.1 alibaba arena runway nvidia togethercompute ollama model-optimization fine-tuning inference-speed video-generation diffusion-models representation-learning local-ai speculative-decoding fp8-quantization context-windows karpathy
Alibaba released compact dense Qwen3-VL models at 4B and 8B sizes with FP8 options, supporting up to 1M context and open vocabulary detection, rivaling larger models like Qwen2.5-VL-72B. Ecosystem support includes MLX-VLM, LM Studio, vLLM, Kaggle models, and Ollama Cloud. In video AI, Arena added Sora 2 models leading in video benchmarks, with Higgsfield Enhancer improving video quality. Runway launched domain-specific workflow apps for creative tasks. Research on Representation Autoencoders for DiTs (RAE-DiT) shows improved diffusion model performance. On local training, NVIDIA DGX Spark enables strong local fine-tuning, while Nanochat by Karpathy offers a minimal stack for training and inference. Together AI introduced ATLAS, a speculative decoding method achieving up to 4× faster inference on DeepSeek-V3.1. These developments highlight advances in efficient model deployment, video AI, local fine-tuning, and inference speed optimization.