Learn about any topic in video form from your favorite personalities
-
Updated
May 24, 2024 - TypeScript
Learn about any topic in video form from your favorite personalities
[ICLR 2024] Cross-Modal Contextualized Diffusion Models for Text-Guided Visual Generation and Editing
Paddle Multimodal Integration and eXploration, supporting mainstream multi-modal tasks, including end-to-end large-scale multi-modal pretrain models and diffusion model toolbox. Equipped with high performance and flexibility.
This repository contains a hand-curated resources for Prompt Engineering with a focus on Generative Pre-trained Transformer (GPT), ChatGPT, PaLM etc
Stable Diffusion, SDXL, LoRA Training, DreamBooth Training, Automatic1111 Web UI, DeepFake, Deep Fakes, TTS, Animation, Text To Video, Tutorials, Guides, Lectures, Courses, ComfyUI, Google Colab, RunPod, NoteBooks, ControlNet, TTS, Voice Cloning, AI, AI News, ML, ML News, News, Tech, Tech News, Kohya LoRA, Kandinsky 2, DeepFloyd IF, Midjourney
[CVPR2024 Highlight] VBench - We Evaluate Video Generation
Ground-A-Video: Zero-shot Grounded Video Editing using Text-to-image Diffusion Models (ICLR 2024)
Script that generates TikTok style videos using ffmpeg, moviepy, chatGPT, and SDXL api within a minute
Cassette is designed to create 30-second explanatory videos suitable for Instagram Reels or YouTube Shorts. Or you may call it a free python alternative to Brainrot.js
VideoCrafter2: Overcoming Data Limitations for High-Quality Video Diffusion Models
VidProM: A Million-scale Real Prompt-Gallery Dataset for Text-to-Video Diffusion Models
Maaagic UI is an open-source UI framework designed to empower developers with seamless integration and advanced features of AI applications.
A GenAI-powered script-to-video converter. Creates beautiful videos from text files. Automatically generates narration, images and audio effects. Can run locally with or without GPUs. This project is experimental in nature, crafted primarily for educational purposes
🔥🔥🔥 A curated list of papers on LLMs-based multimodal generation (image, video, 3D and audio).
MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators
Papers and resources on Controllable Generation using Diffusion Models, including ControlNet, DreamBooth, T2I-Adapter, IP-Adapter.
Implementation of Lumiere, SOTA text-to-video generation from Google Deepmind, in Pytorch
[ICLR 2024] LLM-grounded Video Diffusion Models (LVD): official implementation for the LVD paper
[Arxiv] A Survey on Video Diffusion Models
Implementation of Make-A-Video, new SOTA text to video generator from Meta AI, in Pytorch
Add a description, image, and links to the text-to-video topic page so that developers can more easily learn about it.
To associate your repository with the text-to-video topic, visit your repo's landing page and select "manage topics."