A combination of various deepfake algoritms to quickly create fake audio and video
This project has 4 seperate algorithms.
- first-order-model: a quick deepfake alorithm that generates a video from a base video and driving image
- Speech-Driven Facial Animation: animates a picture to speak an audio input
- Real-Time Voice Cloning Toolbox: a quick text to speech algorithm based off seconds of driving audio
- One-shot Voice Conversion: voice style transform to change the words of one into the words of another
pip install -r requirements.txt
Get from Version Control:
- https://github.com/AliaksandrSiarohin/first-order-model.git
- https://github.com/DinoMan/speech-driven-animation.git
- https://github.com/jjery2243542/adaptive_voice_conversion.git
- https://github.com/jjery2243542/adaptive_voice_conversion.git
and put these into the local project
replace all the dashes in file names with _
Modify each of the files in the following ways:
- go here and download vox-cpk.pth.tar, then place it in first_order_model
- go here and download grid.dat, then replace sda/data/grid.dat
- download: model, delete toolbox/init.py
- download: model and attr, add move them into adaptive_voice_conversion
Next go through each file and correct the imports due to the content root. Go through all the files and add - project_name. - before all necesary imports (if you know a better way, please tell me)