-
Kizdar net |
Kizdar net |
Кыздар Нет
Wan: Open and Advanced Large-Scale Video Generative Models
Feb 25, 2025 · 👍 Multiple Tasks: Wan2.1 excels in Text-to-Video, Image-to-Video, Video Editing, Text-to-Image, and Video-to-Audio, advancing the field of video generation. 👍 Visual Text …
Video-R1: Reinforcing Video Reasoning in MLLMs - GitHub
Feb 23, 2025 · Video-R1 significantly outperforms previous models across most benchmarks. Notably, on VSI-Bench, which focuses on spatial reasoning in videos, Video-R1-7B achieves a …
HunyuanVideo: A Systematic Framework For Large Video
Jan 13, 2025 · We present HunyuanVideo, a novel open-source video foundation model that exhibits performance in video generation that is comparable to, if not superior to, leading …
DepthAnything/Video-Depth-Anything - GitHub
Jan 21, 2025 · This work presents Video Depth Anything based on Depth Anything V2, which can be applied to arbitrarily long videos without compromising quality, consistency, or …
Lightricks/LTX-Video: Official repository for LTX-Video - GitHub
LTX-Video is the first DiT-based video generation model that can generate high-quality videos in real-time. It can generate 30 FPS videos at 1216×704 resolution, faster than it takes to watch …
GitHub - k4yt3x/video2x: A machine learning-based video super ...
A machine learning-based video super resolution and frame interpolation framework. Est. Hack the Valley II, 2018. - k4yt3x/video2x
GitHub - kijai/ComfyUI-WanVideoWrapper
ReCamMaster:. WanVideo2_1_recammaster.mp4. TeaCache (with the old temporary WIP naive version, I2V): Note that with the new version the threshold values should be 10x higher
Lightricks/ComfyUI-LTXVideo: LTX-Video Support for ComfyUI
Sequence Conditioning – Allows motion interpolation from a given frame sequence, enabling video extension from the beginning, end, or middle of the original video. Prompt Enhancer – A …
GitHub - QwenLM/Qwen2.5-VL: Qwen2.5-VL is the multimodal …
In the past five months since Qwen2-VL's release, numerous developers have built new models on the Qwen2-VL vision-language models, providing us with valuable feedback. During this …
【EMNLP 2024 】Video-LLaVA: Learning United Visual ... - GitHub
[2024.09.25] 🔥🔥🔥 Our Video-LLaVA has been accepted at EMNLP 2024! We earn the meta score of 4. [2024.07.27] 🔥🔥🔥 A fine-tuned Video-LLaVA focuses on theme exploration, narrative analysis, …