Cheaper Alternative to Google VEO 3

AI is evolving fast: Fabric 1.0 debuts as the first AI Talking Video Model, Cohere’s Aya 23 boosts multilingual AI, and xAI’s Grok Code Fast 1 speeds up coding workflows.

This Week in AI Big Breakthroughs You Shouldn’t Miss

Here’s a quick look at the latest innovations shaping the AI world:

  • Fabric 1.0 – The world’s first AI Talking Video Model, opening new doors for interactive media.

  • Aya 23 by Cohere – A multilingual model supporting 23 languages with major accuracy gains in reasoning and generation.

  • Grok Code Fast 1 by xAI – A lightning-fast coding assistant built for developers, balancing speed, cost, and transparency.

AI is evolving at an incredible pace, and these launches show just how diverse and impactful the next generation of models will be.

Fabric 1.0: World’s First AI Talking Video Model

Cohere for AI (C4AI), the research arm of Cohere, has announced the open weights release of Aya 23, a new multilingual language model available in 8B and 35B parameter variants that supports 23 widely spoken languages including Arabic, Chinese, Hindi, French, Spanish, and more. Aya 23 offers notable improvements over its predecessor Aya 101 as well as other open models like Google's Gemma and Mistral, with up to 41.6% higher accuracy in multilingual reasoning and superior performance on both discriminative and generative tasks. While the weights of Aya 23 are freely available for researchers to fine-tune and customize under a permissive license, full open sourcing (including training data and model architecture) is not provided. This release is part of Cohere’s ongoing Aya initiative to foster strong, inclusive multilingual AI models, representing a major advancement in bridging global language divides and empowering researchers and practitioners to build diverse language applications.

ByteDance’s Unified AI for Style and Subject Generation

USO (Unified Style and Subject-Driven Generation) by ByteDance is a novel AI model that combines style-driven and subject-driven image generation into a unified framework. It addresses the traditional challenge of balancing subject identity consistency with style fidelity through a two-stage training process involving style-alignment and disentangled content-style learning. By using a large-scale triplet dataset and a style reward learning paradigm, USO achieves high-quality generation with natural, non-plastic portraits while supporting multiple generation modes, including subject-driven, style-driven, and joint style-subject outputs. The project is open-source, providing code, model weights, datasets, and benchmarks to support research and community advancement.

xAI’s Speedy and Affordable AI Coding Model

Grok Code Fast 1 is a new AI reasoning model by xAI designed specifically for agentic coding workflows, offering blazing-fast performance and economical pricing. Built from the ground up with a novel architecture and trained on programming-heavy datasets, it supports multiple programming languages including TypeScript, Python, Java, Rust, C++, and Go. The model excels at typical development tasks such as creating projects from scratch, answering code-related questions, and performing precise bug fixes, all with visible reasoning traces to enhance transparency and steering. It integrates seamlessly with popular IDEs and coding tools like GitHub Copilot and Cursor, delivering a highly responsive experience with tool call optimizations and an extensive 256,000-token context window. Grok Code Fast 1 is currently free on select platforms for a limited time and priced competitively for wide accessibility, striking a balance between speed, cost, and practical coding utility.

Some of you have the same question, “Is it okay to film strangers’ faces, or should I blur them?” Here’s the problem-solving tool, BgBlur, an AI-powered face and background blur that keeps your videos safe, private, and professional.

👉 Try it here: BgBlur

Top AI Products from this week

  • Receiptor AI 2.0 – Forget bookkeeping. Receiptor AI finds every receipt past & present, extracts data in context, auto-categorizes your expenses, and syncs to Xero/QBO. AI easily creates rules and manages your documents. Save time, maximize deductions, stay compliant.

  • Bhava Whether you are a PM, engineer or founder, create flowcharts, visuals, cloud architecture diagrams, swimlane diagrams, UML, ERDs, and much more in seconds.

  • CatDoes CatDoes uses a multi-agent approach to build mobile apps for you and your business using AI. It's so easy to use that anyone, regardless of their technical experience, can build a mobile app and publish it on the App Store using CatDoes. 

  • fileAI MCP - Turn files into usable data instantly. The fileAI MCP server gives AI agents secure access to your files with AI OCR, classification, and schema-based extraction — all through the Model Context Protocol.

  • Copilot Audio Expressions – Copilot Audio Expressions is a free tool that turns text into expressive audio. Use Emotive Mode to direct your own scripts with custom tone and pace, or Story Mode to have Copilot create a full story with narration. All audio is downloadable as MP3.

  • Bugster – Run end-to-end tests in real browsers on every Pull Request — zero maintenance, instant setup, always up-to-date.

  • OAK - OAK is the enterprise AI platform for privacy, security, and control. Empower teams with AI, usage oversight, and data sovereignty. Connect to major LLMs or on-premise LLM hosting with European hosting, max. security, and GDPR compliance.

This week in AI

  • AI BCI Breakthrough - UCLA develops a noninvasive brain-computer interface with an AI co-pilot that boosts speed, accuracy, and independence for paralyzed users.

  • AI Feature Tests - OpenAI trials “effort” control & chat branching in GPT‑5, while Mistral, Gemini, Claude & Perplexity push memory, agents, and automation upgrades.

  • Wan-S2V AI Model - Wan-S2V AI creates film-quality videos from one image + audio, syncing lip movements, expressions & body gestures with cinematic camera work. Supports full/half-body characters.

  • R-4B Auto-Thinking Model - R-4B is an open-source multimodal LLM that adaptively switches between step-by-step reasoning and direct responses, boosting efficiency and state-of-the-art accuracy.

  • MIT’s VaxSeer AI - MIT’s VaxSeer uses ML to predict flu virus evolution and vaccine effectiveness, improving strain selection accuracy and outperforming WHO in most seasons.

Paper of The Day

The Landscape of Agentic Reinforcement Learning for LLMs: A Survey" (arXiv 2509.02547) surveys how agentic reinforcement learning transforms large language models (LLMs) into autonomous decision-making agents. It explores how temporally extended actions, delayed rewards, and goal-directed behavior can be leveraged by LLMs through reinforcement learning frameworks to improve their autonomy and decision-making capabilities in complex environments or tasks. The survey covers the current state-of-the-art methods, challenges, and future directions for integrating agentic reinforcement learning with LLMs to build more robust and effective autonomous agents.

To read the whole paper, go to here.