Gemini 3.1 Flash Image
Gemini 3.1 Flash Image is Google DeepMind’s latest image generation model, combining advanced Pro-level capabilities with lightning-fast performance. It delivers enhanced world knowledge, enabling more accurate subject rendering and data-informed visuals grounded in real-time information. The model improves precision text rendering and in-image translation, making it well-suited for marketing assets, infographics, and localized creative content. Stronger instruction following ensures complex prompts are executed with clarity and accuracy. Gemini 3.1 Flash Image maintains subject consistency across multiple characters and objects within a single workflow. It supports production-ready outputs with customizable aspect ratios and resolutions up to 4K. Available across Gemini, Search, AI Studio, Google Cloud, and more, it brings high-quality visual generation at Flash-level speed.
Learn more
FLUX.2
FLUX.2 is built for real production workflows, delivering high-quality visuals while maintaining character, product, and style consistency across multiple reference images. It handles structured prompts, brand-safe layouts, complex text rendering, and detailed logos with precision. The model supports multi-reference inputs, editing at up to 4 megapixels, and generates both photorealistic scenes and highly stylized compositions. With a focus on reliability, FLUX.2 processes real-world creative tasks—such as infographics, product shots, and UI mockups—with exceptional stability. It represents Black Forest Labs’ open-core approach, pairing frontier-level capability with open-weight models that invite experimentation. Across its variants, FLUX.2 provides flexible options for studios, developers, and researchers who need scalable, customizable visual intelligence.
Learn more
FLUX.1 Krea
FLUX.1 Krea is an open source, guidance-distilled 12 billion-parameter diffusion transformer released by Krea in collaboration with Black Forest Labs, engineered to deliver superior aesthetic control and photorealism while eschewing the generic “AI look.” Fully compatible with the FLUX.1-dev ecosystem, it starts from a raw, untainted base model (flux-dev-raw) rich in world knowledge and employs a two-phase post-training pipeline, supervised fine-tuning on a hand-curated mix of high-quality and synthetic samples, followed by reinforcement learning from human feedback using opinionated preference data, to bias outputs toward a distinct style. By leveraging negative prompts during pre-training, custom loss functions for classifier-free guidance, and targeted preference labels, it achieves significant quality improvements with under one million examples, all without extensive prompting or additional LoRA modules.
Learn more
GLM-Image
GLM-Image is a next-generation, open source image generation model developed by Z.ai, designed to combine deep language understanding with high-fidelity visual synthesis. Unlike traditional diffusion-only models, it uses a hybrid architecture that integrates an autoregressive language model with a diffusion decoder, enabling it to first reason about the structure, meaning, and relationships within a prompt before generating the image itself. This approach allows GLM-Image to excel in scenarios that require precise semantic control, such as generating infographics, presentation slides, posters, and diagrams with accurate embedded text and complex layouts. With a total of around 16 billion parameters, the model achieves strong performance in rendering readable, correctly placed text within images, an area where many image models struggle, while maintaining detailed visual quality and consistency.
Learn more