Explore

Fine-tune FLUX fast

Customize FLUX.1 [dev] with the fast FLUX trainer on Replicate

Train the model to recognize and generate new concepts using a small set of example images, for specific styles, characters, or objects. It's fast (under 2 minutes), cheap (under $2), and gives you a warm, runnable model plus LoRA weights to download.

Official models

Official models are always on, maintained, and have predictable pricing.

View all official models

I want to…

Upscale images

Upscaling models that create high-quality images from low-quality images

Make 3D stuff

Models that generate 3D objects, scenes, radiance fields, textures and multi-views.

Restore images

Models that improve or restore images by deblurring, colorization, and removing noise

Make videos with Wan2.1

Generate videos with Wan2.1, the fastest and highest quality open-source video generation model.

Control image generation

Guide image generation with more than just text. Use edge detection, depth maps, and sketches to get the results you want.

Try for free

Get started with these models without adding a credit card. Whether you're making videos, generating images, or upscaling photos, these are great starting points.

Use official models

Official models are always on, maintained, and have predictable pricing.

Enhance videos

Models that enhance videos with super-resolution, sound effects, motion capture and other useful production effects.

Detect objects

Models that detect or segment objects in images and videos.

Use FLUX fine-tunes

Browse the diverse range of fine-tunes the community has custom-trained on Replicate

Latest models

The model transforms real-life image to Ghibli Style Art Images.

Updated 3.2K runs

WAI-NSFW-illustrious-SDXL v.90

Updated 379 runs

Updated 56 runs

update the official flux-redux-dev to support lora.

Updated 216 runs

DeepCoder-14B-Preview is a code reasoning LLM fine-tuned from DeepSeek-R1-Distilled-Qwen-14B using distributed reinforcement learning (RL) to scale up to long context lengths.

Updated 54 runs

Spanish and English Text to Speech model from Canopy Labs (3b-es_it-ft-research_release)

Updated 104 runs

Updated 4.4K runs

Updated 7 runs

Updated 6.8K runs

This model is an optimised version of stable-diffusion by stability AI that is 3x faster and 3x cheaper.

Updated 163 runs

Updated 58 runs

A 17 billion parameter model with 128 experts

Updated 434.1K runs

A 17 billion parameter model with 16 experts

Updated 233.6K runs

Mediapipe Blendshape Labeler - Predicts the blend shapes of an image.

Updated 206 runs

Fast FLUX DEV -> Flux Controlnet Canny, Controlnet Depth , Controlnet Line Art, Controlnet Upscaler - You can use just one controlnet or All - LORAs: HyperFlex LoRA , Add Details LoRA , Realism LoRA

Updated 167.7K runs

Qwen2.5-Omni is an end-to-end multimodal model designed to perceive diverse modalities, including text, images, audio, and video, while simultaneously generating text and natural speech responses in a streaming manner.

Updated 2.8K runs

Updated 41 runs

Controllable generative AI art

Updated 572 runs

Updated 2.9K runs

Generates realistic talking face animations from a portrait image and audio using the CVPR 2025 Sonic model

Updated 9.5K runs

Transform your portrait photos into any style or setting while preserving your facial identity

Updated 7.8K runs

Wan 2.1 1.3b Video to Video. Wan is a powerful visual generation model developed by Tongyi Lab of Alibaba Group

Updated 777 runs

Easily create video datasets with auto-captioning for Hunyuan-Video LoRA finetuning

Updated 618 runs

Cost-optimized MMAudio V2 (T4 GPU): Add sound to video using this version running on T4 hardware for lower cost. Synthesizes high-quality audio from video content.

Updated 506 runs

Add sound to video using the MMAudio V2 model. An advanced AI model that synthesizes high-quality audio from video content, enabling seamless video-to-audio transformation.

Updated 697.4K runs

A redux adapter trained from scratch on Flex.1-alpha, that also works with FLUX.1-dev

Updated 209 runs

Indic Parler-TTS Pretrained is a multilingual Indic extension of Parler-TTS Mini.

Updated 49 runs

SANA-Sprint: One-Step Diffusion with Continuous-Time Consistency Distillation

Updated 387K runs

flux_schnell model img2img inference

Updated 139.6K runs

Open-weight inpainting model for editing and extending images. Guidance-distilled from FLUX.1 Fill [pro].

Updated 540.2K runs

FLUX1.1 [pro] in ultra and raw modes. Images are up to 4 megapixels. Use raw mode for realism.

Updated 14.1M runs

Faster, better FLUX Pro. Text-to-image model with excellent image quality, prompt adherence, and output diversity.

Updated 39.9M runs

State-of-the-art image generation with top of the line prompt following, visual quality, image detail and output diversity.

Updated 12.1M runs

Professional inpainting and outpainting model with state-of-the-art performance. Edit or extend images with natural, seamless results.

Updated 1.8M runs

Professional edge-guided image generation. Control structure and composition using Canny edge detection

Updated 292.2K runs

Simple binary sentiment analysis with BERT

Updated 56 runs

TripoSG unofficial implementation

Updated 270 runs

Updated 3K runs

Updated 2.2K runs

Updated 52 runs

For the paper "Structured 3D Latents for Scalable and Versatile 3D Generation".

Updated 280 runs

A model Flux.1-dev-Controlnet-Upscaler by www.androcoders.in

Updated 108 runs

Accelerated inference for Wan 2.1 14B text to video, a comprehensive and open suite of video foundation models that pushes the boundaries of video generation.

Updated 140.8K runs

Accelerated inference for Wan 2.1 14B text to video with high resolution, a comprehensive and open suite of video foundation models that pushes the boundaries of video generation.

Updated 32.8K runs

Accelerated inference for Wan 2.1 14B image to video, a comprehensive and open suite of video foundation models that pushes the boundaries of video generation.

Updated 324.8K runs

Accelerated inference for Wan 2.1 14B image to video with high resolution, a comprehensive and open suite of video foundation models that pushes the boundaries of video generation.

Updated 69K runs

Updated 14 runs

Virtual fitting of clothes

Updated 2K runs

Updated 1.3K runs

YUMEMONO

Updated 41 runs