Mistral Medium 3.1
Mistral Medium 3.1 is the latest frontier-class multimodal foundation model released in August 2025, designed to deliver advanced reasoning, coding, and multimodal capabilities while dramatically reducing deployment complexity and costs. It builds on the highly efficient architecture of Mistral Medium 3, renowned for offering state-of-the-art performance at up to 8-times lower cost than leading large models, enhancing tone consistency, responsiveness, and accuracy across diverse tasks and modalities. The model supports deployment across hybrid environments, on-premises systems, and virtual private clouds, and it achieves competitive performance relative to high-end models such as Claude Sonnet 3.7, Llama 4 Maverick, and Cohere Command A. Ideal for professional and enterprise use cases, Mistral Medium 3.1 excels in coding, STEM reasoning, language understanding, and multimodal comprehension, while maintaining broad compatibility with custom workflows and infrastructure.
Learn more
Google Cloud Vision AI
Derive insights from your images in the cloud or at the edge with AutoML Vision or use pre-trained Vision API models to detect emotion, understand text, and more. Google Cloud offers two computer vision products that use machine learning to help you understand your images with industry-leading prediction accuracy. Automate the training of your own custom machine learning models. Simply upload images and train custom image models with AutoML Vision’s easy-to-use graphical interface; optimize your models for accuracy, latency, and size; and export them to your application in the cloud, or to an array of devices at the edge. Google Cloud’s Vision API offers powerful pre-trained machine learning models through REST and RPC APIs. Assign labels to images and quickly classify them into millions of predefined categories. Detect objects and faces, read printed and handwritten text, and build valuable metadata into your image catalog.
Learn more
Command A Reasoning
Command A Reasoning is Cohere’s most advanced enterprise-ready language model, engineered for high-stakes reasoning tasks and seamless integration into AI agent workflows. The model delivers exceptional reasoning performance, efficiency, and controllability, scaling across multi-GPU setups with support for up to 256,000-token context windows, ideal for handling long documents and multi-step agentic tasks. Organizations can fine-tune output precision and latency through a token budget, allowing a single model to flexibly serve both high-accuracy and high-throughput use cases. It powers Cohere’s North platform with leading benchmark performance and excels in multilingual contexts across 23 languages. Designed with enterprise safety in mind, it balances helpfulness with robust safeguards against harmful outputs. A lightweight deployment option allows running the model securely on a single H100 or A100 GPU, simplifying private, scalable use.
Learn more
Ray2
Ray2 is a large-scale video generative model capable of creating realistic visuals with natural, coherent motion. It has a strong understanding of text instructions and can take images and video as input. Ray2 exhibits advanced capabilities as a result of being trained on Luma’s new multi-modal architecture scaled to 10x compute of Ray1. Ray2 marks the beginning of a new generation of video models capable of producing fast coherent motion, ultra-realistic details, and logical event sequences. This increases the success rate of usable generations and makes videos generated by Ray2 substantially more production-ready. Text-to-video generation is available in Ray2 now, with image-to-video, video-to-video, and editing capabilities coming soon. Ray2 brings a whole new level of motion fidelity. Smooth, cinematic, and jaw-dropping, transform your vision into reality. Tell your story with stunning, cinematic visuals. Ray2 lets you craft breathtaking scenes with precise camera movements.
Learn more