Explore

I want to…

Restore images

Models that improve or restore images by deblurring, colorization, and removing noise

Enhance videos

Models that enhance videos with super-resolution, sound effects, motion capture and other useful production effects.

Detect objects

Models that detect or segment objects in images and videos.

Make 3D stuff

Models that generate 3D objects, scenes, radiance fields, textures and multi-views.

Use FLUX fine-tunes

Browse the diverse range of fine-tunes the community has custom-trained on Replicate

Control image generation

Guide image generation with more than just text. Use edge detection, depth maps, and sketches to get the results you want.

Latest models

Updated 46 runs

welcome to contact us. youkpan@gmail.com

Updated 100 runs

Updated 14 runs

Updated 43 runs

wan-video/wan-2.1-1.3b

Generate 5s 480p videos. Wan is an advanced and powerful visual generation model developed by Tongyi Lab of Alibaba Group

Updated 1.9K runs

wavespeedai/wan-2.1-i2v-480p

Accelerated inference for Wan 2.1 14B image-to-video, a comprehensive and open suite of video foundation models that pushes the boundaries of video generation.

Updated 910 runs

wavespeedai/wan-2.1-i2v-720p

Accelerated inference for Wan 2.1 14B image-to-video with high resolution, a comprehensive and open suite of video foundation models that pushes the boundaries of video generation.

Updated 514 runs

wavespeedai/wan-2.1-t2v-480p

Accelerated inference for Wan 2.1 14B, a comprehensive and open suite of video foundation models that pushes the boundaries of video generation.

Updated 845 runs

wavespeedai/wan-2.1-t2v-720p

Accelerated inference for Wan 2.1 14B with high resolution, a comprehensive and open suite of video foundation models that pushes the boundaries of video generation.

Updated 211 runs

ideogram-ai/ideogram-v2a-turbo

Like Ideogram v2 turbo, but now faster and cheaper

Updated 587 runs

ideogram-ai/ideogram-v2a

Like Ideogram v2, but faster and cheaper

Updated 1.2K runs

Updated 53 runs

Updated 106 runs

Updated 32 runs

ibm-granite/granite-vision-3.2-2b

Granite-Vision-3.2-2B is a compact and efficient vision-language model, specifically designed for visual document understanding.

Updated 1.3K runs

Updated 2 runs

Updated 21 runs

Removes furniture

Updated 162 runs

In-Context LoRA with Image-to-Image and Inpainting to apply your logo to anything

Updated 1.7K runs

anthropic/claude-3.7-sonnet

The most intelligent Claude model and the first hybrid reasoning model on the market (claude-3-7-sonnet-20250219)

Updated 1.4K runs

Generate high-quality videos from text prompts using StepVideo

Updated 48 runs

minimax/video-01-director

Generate videos with specific camera movements

Updated 3.4K runs

Updated 417 runs

Updated 138 runs

Updated 3 runs

Updated 14 runs

Updated 228 runs

Updated 2 runs

This model generates pose variation of a cartoon character. It preserves the cartoon identity. Use this model to augment training dataset for any cartoon character created through AI. The augmented dataset can be used to train a LoRA model.

Updated 2.4K runs

Realistic text-to-image by TiwazM

Updated 1K runs

Updated 233 runs

Updated 52 runs

Updated 60 runs

Updated 2K runs

Updated 294 runs

Updated 25 runs

Sa2VA: Marrying SAM2 with LLaVA for Dense Grounded Understanding of Images and Videos

Updated 278 runs

Sa2VA: Marrying SAM2 with LLaVA for Dense Grounded Understanding of Images and Videos

Updated 56 runs

Sa2VA: Marrying SAM2 with LLaVA for Dense Grounded Understanding of Images and Videos

Updated 19 runs

Sa2VA: Marrying SAM2 with LLaVA for Dense Grounded Understanding of Images and Videos

Updated 38 runs

Sa2VA: Marrying SAM2 with LLaVA for Dense Grounded Understanding of Images and Videos

Updated 15 runs

Sa2VA: Marrying SAM2 with LLaVA for Dense Grounded Understanding of Images and Videos

Updated 149 runs

Updated 544 runs

A version of the DeepSeek-R1 model that has been post trained to provide unbiased, accurate, and factual information by Perplexity

Updated 42 runs

Updated 9 runs

wavespeedai/step-video

Accelerated inference for Step-Video-T2V, a state-of-the-art (SoTA) text-to-video pre-trained model with 30 billion parameters and the capability to generate videos up to 204 frames.

Updated 123 runs

An upscaler based on tile and inpaint controlnets, aimed to preserve the original image while injecting more details.

Updated 50 runs