Explore

I want to…

Restore images

Models that improve or restore images by deblurring, colorization, and removing noise

Upscale images

Upscaling models that create high-quality images from low-quality images

Make 3D stuff

Models that generate 3D objects, scenes, radiance fields, textures and multi-views.

Latest models

Pre-trained Subject Representation for Controllable Text-to-Image Generation and Editing with ControlNet

Updated 163 runs

incrediblye fast whisper using openai/whisper-large-v3 NOT the distil model

Updated 126 runs

Guiding Instruction-based Image Editing via Multimodal Large Language Models

Updated 5.7K runs

A capable large language model for natural language to SQL generation.

Updated 21.3K runs

Depth Anything on full video files

Updated 327 runs

Updated 174 runs

Hand Refiner 512x512

Updated 1.2K runs

Updated 378 runs

Yuan2.0 is a new generation LLM developed by IEIT System, enhanced the model's understanding of semantics, mathematics, reasoning, code, knowledge, and other aspects.

Updated 38 runs

LGM: Large Multi-View Gaussian Model for High-Resolution 3D Content Creation

Updated 3.6K runs

phixtral-2x2_8 is the first Mixure of Experts (MoE) made with two microsoft/phi-2 models, inspired by the mistralai/Mixtral-8x7B-v0.1 architecture

Updated 212 runs

Updated 572 runs

BGE-M3, the first embedding model which supports multiple retrieval mode, multilingual and multi-granularity retrieval.

Updated 226 runs

MetaVoice-1B: 1.2B parameter base model trained on 100K hours of speech

Updated 10K runs

Remove background from image

Updated 12.9K runs

incredibly fast whisper using openai/whisper-medium.en NOT the distil model

Updated 246 runs

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

Updated 3.8K runs

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

Updated 672 runs

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

Updated 210 runs

Background removal model developed by BRIA.AI, trained on a carefully selected dataset and is available as an open-source model for non-commercial use.

Updated 46K runs

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

Updated 19 runs

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

Updated 76 runs

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

Updated 58 runs

Updated 174 runs

Remove background from images using BRIA-RMBG-1.4

Updated 10.2K runs

NeverSleep's MiquMaid v1 70B Miqu Finetune, GGUF Q3_K_M quantized by NeverSleep.

Updated 14.1K runs

Base version of Mamba 2.8B, a 2.8 billion parameter state space language model

Updated 230 runs

Base version of Mamba 130M, a 130 million parameter state space language model

Updated 110 runs

Base version of Mamba 370M, a 370 million parameter state space language model

Updated 41 runs

Base version of Mamba 790M, a 790 million parameter state space language model

Updated 43 runs

Base version of Mamba 2.8B Slim Pyjama, a 2.8 billion parameter state space language model

Updated 51 runs

Base version of Mamba 1.4B, a 1.4 billion parameter state space language model

Updated 59 runs

Create a video from an image

Updated 15.2K runs

Merge two images, with an optional third for controlnet.

Updated 4.3K runs

this is a first model

Updated 55 runs

A Visual Language Model for GUI Agents

Updated 2.2K runs

Bokeh Prediction, a hybrid bokeh rendering framework that combines a neural renderer with a classical approach. It generates high-resolution, adjustable bokeh effects from a single image and potentially imperfect disparity maps.

Updated 472 runs

AnimateLCM Cartoon3D Model

Updated 1.3K runs

Finetuned E5 embeddings for instruct based on Mistral.

Updated 131 runs

MoE-LLaVA

Updated 1.4M runs

Updated 127 runs

LLaVA v1.6: Large Language and Vision Assistant (Nous-Hermes-2-34B)

Updated 1.6M runs

LLaVA v1.6: Large Language and Vision Assistant (Vicuna-13B)

Updated 3.3M runs

LLaVA v1.6: Large Language and Vision Assistant (Vicuna-7B)

Updated 61.6K runs

LLaVA v1.6: Large Language and Vision Assistant (Mistral-7B)

Updated 4.8M runs

one-shot-talking-face-replicate

Updated 1.4K runs

unet clothing segment

Updated 727 runs

Updated 31 runs

Yi-VL-34B is the first open-source 34B VL model worldwide. It demonstrates exceptional performance, ranking first among all existing open-source models in the latest benchmarks including MMMU and CMMMU.

Updated 292 runs

🖼️ Super fast 1.5B Image Captioning/VQA Multimodal LLM (Image-to-Text) 🖋️

Updated 2K runs