Stars
Generate Diffuse Textures on Meshes directly in Blender 3D with Stable Diffusion.
Finetune Llama 3.3, DeepSeek-R1 & Reasoning LLMs 2x faster with 70% less memory! 🦥
Build AI-powered applications with React, Svelte, Vue, and Solid
A set of beautifully-designed, accessible components and a code distribution platform. Works with your favorite frameworks. Open Source. Open Code.
An opinionated collection of components, hooks, and utilities for your Next.js project.
MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators
Official implementations for paper: Dynamic Typography: Bringing Text to Life via Video Diffusion Prior
Official code for the CVPR 2025 paper "SemanticDraw: Towards Real-Time Interactive Content Creation from Image Diffusion Models."
Lifting ControlNet for Generalized Depth Conditioning
Tiled Diffusion and VAE optimize, licensed under CC BY-NC-SA 4.0
A growing curation of Text-to-3D, Diffusion-to-3D works.
A simple canvas drawing web app with responsive UI. Made with TypeScript, React, and Next.js.
Python library for video editing, presentation video generation, motion graphics, shader art coding, and other video production tasks
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
Create a draggable and resizable dashboard in Streamlit, featuring Material UI widgets, Monaco editor (Visual Studio Code), Nivo charts, and more!
[ECCV 2022] Meta-Sampler: Almost Universal yet Task-Oriented Sampling for Point Clouds
High-Resolution Image Synthesis with Latent Diffusion Models
[AAAI 2022] Pose Adaptive Dual Mixup for Few-Shot Single-View 3D Reconstruction
Learn by experimenting on state-of-the-art machine learning models and algorithms with Jupyter Notebooks.
Try out deep learning models online on Google Colab
TikTok Scraper. Download video posts, collect user/trend/hashtag/music feed metadata, sign URL and etc.
Official implementation of the paper WAV2CLIP: LEARNING ROBUST AUDIO REPRESENTATIONS FROM CLIP
Google Research
Source code for models described in the paper "AudioCLIP: Extending CLIP to Image, Text and Audio" (https://arxiv.org/abs/2106.13043)
Experimental CartoonGAN (Chen et.al.) implementation for quicker background generation for posters and new episodes