The Wayback Machine - https://web.archive.org/web/20231210122541/https://github.com/Flode-Labs/vid2densepose
Skip to content

Flode-Labs/vid2densepose

main
Switch branches/tags

Name already in use

A tag already exists with the provided branch name. Many Git commands accept both tag and branch names, so creating this branch may cause unexpected behavior. Are you sure you want to create this branch?
Code

Files

Permalink
Failed to load latest commit information.
Type
Name
Latest commit message
Commit time
December 7, 2023 23:17
December 5, 2023 18:48
December 9, 2023 23:07
December 5, 2023 19:54

Vid2DensePose

Open In Colab

Overview

The Vid2DensePose is a powerful tool designed for applying the DensePose model to videos, generating detailed "Part Index" visualizations for each frame. This tool is exceptionally useful for enhancing animations, particularly when used in conjunction with MagicAnimate for temporally consistent human image animation.

Key Features

  • Enhanced Output: Produces video files showcasing DensePosedata in a vivid, color-coded format.
  • MagicAnimate Integration: Seamlessly compatible with MagicAnimate to foster advanced human animation projects.

Prerequisites

To utilize this tool, ensure the installation of:

  • Python 3.8 or later
  • PyTorch (preferably with CUDA for GPU support)
  • Detectron2

Installation Steps

  1. Clone the repository:

    git clone https://github.com/Flode-Labs/vid2densepose.git
    cd vid2densepose
  2. Install necessary Python packages:

    pip install -r requirements.txt
  3. Clone the Detectron repository:

    git clone https://github.com/facebookresearch/detectron2.git

Usage Guide

Run the script:

python main.py -i sample_videos/input_video.mp4 -o sample_videos/output_video.mp4

The script processes the input video and generates an output with the densePose format.

Gradio version

You can also use the Gradio to run the script with an interface. To do so, run the following command:

python app.py

Integration with MagicAnimate

For integration with MagicAnimate:

  1. Create the densepose video using the steps outlined above.
  2. Use this output as an input to MagicAnimate for generating temporally consistent animations.

Acknowledgments

Special thanks to:

  • Facebook AI Research (FAIR) for the development of DensePose.
  • The contributors of the Detectron2 project.
  • Gonzalo Vidal for the sample videos.
  • Sylvain Filoni for the deployment of the Gradio Space in Hugging Face.

Support

For any inquiries or support, please file an issue in our GitHub repository's issue tracker.

About

Convert your videos to densepose and use it on MagicAnimate

Topics

Resources

License

Stars

Watchers

Forks

Contributors 4

  •  
  •  
  •  
  •  

Languages