Explore
Featured models
minimax / video-01
Generate 6s videos with prompts or images. (Also known as Hailuo)
black-forest-labs / flux-fill-pro
Professional inpainting and outpainting model with state-of-the-art performance. Edit or extend images with natural, seamless results.
black-forest-labs / flux-1.1-pro-ultra
FLUX1.1 [pro] in ultra and raw modes. Images are up to 4 megapixels. Use raw mode for realism.
black-forest-labs / flux-redux-dev
Open-weight image variation model. Create new versions while preserving key elements of your original.
recraft-ai / recraft-v3
Recraft V3 (code-named red_panda) is a text-to-image model with the ability to generate long texts, and images in a wide list of styles. As of today, it is SOTA in image generation, proven by the Text-to-Image Benchmark by Artificial Analysis
davisbrown / flux-half-illustration
Flux lora, use "in the style of TOK" to trigger generation, creates half photo half illustrated elements
I want to…
Generate images
Models that generate images from text prompts
Use a language model
Models that can understand and generate text
Upscale images
Upscaling models that create high-quality images from low-quality images
Caption images
Models that generate text from images
The FLUX family of models
The FLUX family of text-to-image models from Black Forest Labs
Restore images
Models that improve or restore images by deblurring, colorization, and removing noise
Get embeddings
Models that generate embeddings from inputs
Extract text from images
Optical character recognition (OCR) and text extraction
Transcribe speech
Models that convert speech to text
Use handy tools
Toolbelt-type models for videos and images.
Chat with images
Ask language models about images
Edit images
Tools for manipulating images.
Use a face to make images
Make realistic images of people instantly
Flux fine-tunes
Browse the diverse range of fine-tunes the community has custom-trained on Replicate
Generate music
Models to generate and modify music
Generate videos
Models that create and edit videos
Generate speech
Convert text to speech
Make 3D stuff
Models that generate 3D objects, scenes, radiance fields, textures and multi-views.
Get structured data
Language models that support grammar-based decoding as well as jsonschema constraints.
Popular models
SDXL-Lightning by ByteDance: a fast text-to-image model that makes high-quality images in 4 steps
Fine-Tuned Vision Transformer (ViT) for NSFW Image Classification
A text-to-image generative AI model that creates beautiful images
Real-ESRGAN with optional face correction and adjustable upscale
Latest models
Segment foreground objects with high resolution and matting, using InSPyReNet
Three models in one Cog: Absolute Reality v1.8.1, DreamShaper v8 and Meina V4
Source: gradientai/Llama-3-8B-Instruct-Gradient-4194k ✦ Quant: solidrust/Llama-3-8B-Instruct-Gradient-4194k-AWQ ✦ Extending LLama-3 8B's context length from 8k to 4194K
CLIP Interrogator for SDXL optimizes text prompts to match a given image
📖 PuLID: Pure and Lightning ID Customization via Contrastive Alignment
PaliGemma 3B, an open VLM by Google, pre-trained with 224*224 input images and 128 token input/output text sequences
A model which generates text in response to an input image and prompt.
Generate image with transparent background
Yi-1.5 is continuously pre-trained on Yi with a high-quality corpus of 500B tokens and fine-tuned on 3M diverse fine-tuning samples
InstantMesh: Efficient 3D Mesh Generation from a Single Image with Sparse-view LRMs
Blip 3 / XGen-MM, Answers questions about images ({blip3,xgen-mm}-phi3-mini-base-r-v1)
Dolphin is uncensored. I have filtered the dataset to remove alignment and bias. This makes the model more compliant.
return CLIP features for the dfn5b-clip-vit-h-14-384, current highest average perf. in openclip models leaderboard.
Dolphin is uncensored. I have filtered the dataset to remove alignment and bias. This makes the model more compliant.
Dolphin is uncensored. I have filtered the dataset to remove alignment and bias. This makes the model more compliant.
BLIP3(XGen-MM) is a series of foundational Large Multimodal Models (LMMs) developed by Salesforce AI Research
Transcribe audios using OpenAI's Whisper with stabilizing timestamps by stable-ts python package.
Use a face to instantly make images. Uses SDXL Lightning checkpoints.
Cog to turn minimally-formatted plaintext into pdfs (using tex on the backend)
Dark Sushi Mix 2.25D Model with vae-ft-mse-840000-ema (Text2Img, Img2Img and Inpainting)
DeepSeek LLM, an advanced language model comprising 67 billion parameters. Trained from scratch on a vast dataset of 2 trillion tokens in both English and Chinese
A llama-3 based moderation and safeguarding language model
InstantID. ControlNets. More base SDXL models. And the latest ByteDance's ⚡️SDXL-Lightning !⚡️
The img2img pipeline that makes an anime-style image of a person. It uses one of sd1.5 models as a base, depth-estimation as a ControleNet and IPadapter model for face consistency.
Consistent Self-Attention for Long-Range Image and Video Generation
StoryDiffusion: Consistent Self-Attention for Long-Range Image and Video Generation
Robust face restoration algorithm for old photos / AI-generated faces (adapted to work with video inputs)