Skip to content

Commit 92db20e

Browse files
authored
Merge pull request #632 from KRSHH/main
Unreverting New UI after Fixes, Unreverted README
2 parents 92a0994 + f1e3657 commit 92db20e

File tree

2 files changed

+961
-340
lines changed

2 files changed

+961
-340
lines changed

README.md

+120-97
Original file line numberDiff line numberDiff line change
@@ -1,173 +1,186 @@
11

22
![demo-gif](demo.gif)
3+
![demo-gif](avgpcperformancedemo.gif)
34

5+
## Deep Live Cam
6+
7+
Real-time face swap and video deepfake with a single click and only a single image.
48

59
## Disclaimer
6-
This software is meant to be a productive contribution to the rapidly growing AI-generated media industry. It will help artists with tasks such as animating a custom character or using the character as a model for clothing etc.
710

8-
The developers of this software are aware of its possible unethical applications and are committed to take preventative measures against them. It has a built-in check which prevents the program from working on inappropriate media including but not limited to nudity, graphic content, sensitive material such as war footage etc. We will continue to develop this project in the positive direction while adhering to law and ethics. This project may be shut down or include watermarks on the output if requested by law.
11+
This software is intended as a productive contribution to the AI-generated media industry. It aims to assist artists with tasks like animating custom characters or using them as models for clothing, etc.
12+
13+
We are aware of the potential for unethical applications and are committed to preventative measures. A built-in check prevents the program from processing inappropriate media (nudity, graphic content, sensitive material like war footage, etc.). We will continue to develop this project responsibly, adhering to law and ethics. We may shut down the project or add watermarks if legally required.
14+
15+
Users are expected to use this software responsibly and legally. If using a real person's face, obtain their consent and clearly label any output as a deepfake when sharing online. We are not responsible for end-user actions.
916

10-
Users of this software are expected to use this software responsibly while abiding by local laws. If the face of a real person is being used, users are required to get consent from the concerned person and clearly mention that it is a deepfake when posting content online. Developers of this software will not be responsible for actions of end-users.
1117

12-
## New Features
18+
## Features
19+
1320
### Resizable Preview Window
1421

15-
Dynamically improve the performance by using the `--live-resizable` parameter
22+
Dynamically improve performance using the `--live-resizable` parameter.
23+
1624
![resizable-gif](resizable.gif)
1725

1826
### Face Mapping
1927

20-
Track faces and change it on the fly
28+
Track and change faces on the fly.
2129

2230
![face_mapping_source](face_mapping_source.gif)
2331

24-
source video
32+
**Source Video:**
2533

2634
![face-mapping](face_mapping.png)
2735

28-
Tick this switch
36+
**Enable Face Mapping:**
2937

3038
![face-mapping2](face_mapping2.png)
3139

32-
Map the faces
40+
**Map the Faces:**
3341

3442
![face_mapping_result](face_mapping_result.gif)
3543

36-
And see the magic!
44+
**See the Magic!**
3745

46+
## Quick Start (Windows / Nvidia)
3847

39-
## Want to skip the installation and just run it?
40-
<details>
41-
Here's the link without the tedious installation script below
48+
[Download pre-built version with CUDA support](https://hacksider.gumroad.com/l/vccdmm)
4249

43-
[Windows / Nvidia](https://hacksider.gumroad.com/l/vccdmm) CUDA still required
44-
</details>
45-
46-
## How do I install it?
50+
## Installation (Manual)
4751

52+
### Basic Installation (CPU)
4853

49-
### Basic: It is more likely to work on your computer but it will also be very slow. You can follow instructions for the basic install (This usually runs via **CPU**)
50-
#### 1.Setup your platform
51-
- python (3.10 recommended)
52-
- pip
53-
- git
54-
- [ffmpeg](https://www.youtube.com/watch?v=OlNWCpFdVMA)
55-
- [visual studio 2022 runtimes (windows)](https://visualstudio.microsoft.com/visual-cpp-build-tools/)
56-
#### 2. Clone Repository
57-
https://github.com/hacksider/Deep-Live-Cam.git
54+
This is more likely to work on your computer but will be slower as it utilizes the CPU.
5855

59-
#### 3. Download Models
56+
**1. Setup Your Platform**
6057

61-
1. [GFPGANv1.4](https://huggingface.co/hacksider/deep-live-cam/resolve/main/GFPGANv1.4.pth)
62-
2. [inswapper_128.onnx](https://huggingface.co/hacksider/deep-live-cam/resolve/main/inswapper_128.onnx) *(Note: Use this [replacement version](https://github.com/facefusion/facefusion-assets/releases/download/models/inswapper_128.onnx) if an issue occurs on your computer)*
58+
- Python (3.10 recommended)
59+
- pip
60+
- git
61+
- [ffmpeg](https://www.youtube.com/watch?v=OlNWCpFdVMA)
62+
- [Visual Studio 2022 Runtimes (Windows)](https://visualstudio.microsoft.com/visual-cpp-build-tools/)
6363

64-
Then put those 2 files on the "**models**" folder
64+
**2. Clone Repository**
6565

66-
#### 4. Install dependency
67-
We highly recommend to work with a `venv` to avoid issues.
66+
```bash
67+
https://github.com/hacksider/Deep-Live-Cam.git
6868
```
69+
70+
**3. Download Models**
71+
72+
1. [GFPGANv1.4](https://huggingface.co/hacksider/deep-live-cam/resolve/main/GFPGANv1.4.pth)
73+
2. [inswapper_128_fp16.onnx](https://huggingface.co/hacksider/deep-live-cam/resolve/main/inswapper_128.onnx) (Note: Use this [replacement version](https://github.com/facefusion/facefusion-assets/releases/download/models/inswapper_128.onnx) if you encounter issues)
74+
75+
Place these files in the "**models**" folder.
76+
77+
**4. Install Dependencies**
78+
79+
We highly recommend using a `venv` to avoid issues.
80+
81+
```bash
6982
pip install -r requirements.txt
7083
```
71-
For MAC OS, You have to install or upgrade python-tk package:
72-
```
84+
85+
**For macOS:** Install or upgrade the `python-tk` package:
86+
87+
```bash
7388
brew install [email protected]
7489
```
75-
##### DONE!!! If you don't have any GPU, You should be able to run Deep-Live-Cam using `python run.py` command. Keep in mind that while running the program for first time, it will download some models which can take time depending on your network connection.
7690

77-
#### 5. Proceed if you want to use GPU acceleration (optional)
91+
**Run:** If you don't have a GPU, you can run Deep-Live-Cam using `python run.py`. Note that initial execution will download models (~300MB).
92+
93+
94+
### GPU Acceleration (Optional)
7895

7996
<details>
8097
<summary>Click to see the details</summary>
8198

82-
### CUDA Execution Provider (Nvidia)*
99+
**CUDA Execution Provider (Nvidia)**
83100

84-
1. Install [CUDA Toolkit 11.8](https://developer.nvidia.com/cuda-11-8-0-download-archive)
85-
2. Install dependencies:
86-
```
101+
1. Install [CUDA Toolkit 11.8](https://developer.nvidia.com/cuda-11-8-0-download-archive)
102+
2. Install dependencies:
103+
```bash
87104
pip uninstall onnxruntime onnxruntime-gpu
88105
pip install onnxruntime-gpu==1.16.3
89106
```
90-
91-
3. Usage in case the provider is available:
92-
```
107+
3. Usage:
108+
```bash
93109
python run.py --execution-provider cuda
94110
```
95111

96-
### [](https://github.com/s0md3v/roop/wiki/2.-Acceleration#coreml-execution-provider-apple-silicon)CoreML Execution Provider (Apple Silicon)
112+
**CoreML Execution Provider (Apple Silicon)**
97113

98-
1. Install dependencies:
99-
100-
```
114+
1. Install dependencies:
115+
```bash
101116
pip uninstall onnxruntime onnxruntime-silicon
102117
pip install onnxruntime-silicon==1.13.1
103118
```
104-
105-
2. Usage in case the provider is available:
106-
```
119+
2. Usage:
120+
```bash
107121
python run.py --execution-provider coreml
108122
```
109123

110-
### [](https://github.com/s0md3v/roop/wiki/2.-Acceleration#coreml-execution-provider-apple-legacy)CoreML Execution Provider (Apple Legacy)
124+
**CoreML Execution Provider (Apple Legacy)**
111125

112-
1. Install dependencies:
113-
```
126+
1. Install dependencies:
127+
```bash
114128
pip uninstall onnxruntime onnxruntime-coreml
115129
pip install onnxruntime-coreml==1.13.1
116130
```
117-
118-
2. Usage in case the provider is available:
119-
```
131+
2. Usage:
132+
```bash
120133
python run.py --execution-provider coreml
121134
```
122135

123-
### [](https://github.com/s0md3v/roop/wiki/2.-Acceleration#directml-execution-provider-windows)DirectML Execution Provider (Windows)
136+
**DirectML Execution Provider (Windows)**
124137

125-
1. Install dependencies:
126-
```
138+
1. Install dependencies:
139+
```bash
127140
pip uninstall onnxruntime onnxruntime-directml
128141
pip install onnxruntime-directml==1.15.1
129142
```
130-
131-
2. Usage in case the provider is available:
132-
```
143+
2. Usage:
144+
```bash
133145
python run.py --execution-provider directml
134146
```
135147

136-
### [](https://github.com/s0md3v/roop/wiki/2.-Acceleration#openvino-execution-provider-intel)OpenVINO™ Execution Provider (Intel)
148+
**OpenVINO™ Execution Provider (Intel)**
137149

138-
1. Install dependencies:
139-
```
150+
1. Install dependencies:
151+
```bash
140152
pip uninstall onnxruntime onnxruntime-openvino
141153
pip install onnxruntime-openvino==1.15.0
142154
```
143-
144-
2. Usage in case the provider is available:
145-
```
155+
2. Usage:
156+
```bash
146157
python run.py --execution-provider openvino
147158
```
159+
148160
</details>
149161

150-
## How do I use it?
151-
> Note: When you run this program for the first time, it will download some models ~300MB in size.
152162

153-
Executing `python run.py` command will launch this window:
154-
![gui-demo](instruction.png)
163+
## Usage
155164

156-
Choose a face (image with desired face) and the target image/video (image/video in which you want to replace the face) and click on `Start`. Open file explorer and navigate to the directory you select your output to be in. You will find a directory named `<video_title>` where you can see the frames being swapped in realtime. Once the processing is done, it will create the output file. That's it.
165+
**1. Image/Video Mode**
157166

158-
## For the webcam mode
159-
Just follow the clicks on the screenshot
160-
1. Select a face
161-
2. Click live
162-
3. Wait for a few seconds (it takes a longer time, usually 10 to 30 seconds before the preview shows up)
167+
- Execute `python run.py`.
168+
- Choose a source face image and a target image/video.
169+
- Click "Start".
170+
- The output will be saved in a directory named after the target video.
163171

164-
![demo-gif](demo.gif)
172+
**2. Webcam Mode**
165173

166-
Just use your favorite screencapture to stream like OBS
167-
> Note: In case you want to change your face, just select another picture, the preview mode will then restart (so just wait a bit).
174+
- Execute `python run.py`.
175+
- Select a source face image.
176+
- Click "Live".
177+
- Wait for the preview to appear (10-30 seconds).
178+
- Use a screen capture tool like OBS to stream.
179+
- To change the face, select a new source image.
168180

181+
![demo-gif](demo.gif)
169182

170-
Additional command line arguments are given below. To learn out what they do, check [this guide](https://github.com/s0md3v/roop/wiki/Advanced-Options).
183+
## Command Line Arguments
171184

172185
```
173186
options:
@@ -194,7 +207,8 @@ options:
194207

195208
Looking for a CLI mode? Using the -s/--source argument will make the run program in cli mode.
196209

197-
### Webcam mode on Windows 11 using WSL2 Ubuntu (optional)
210+
211+
## Webcam Mode on WSL2 Ubuntu (Optional)
198212

199213
<details>
200214
<summary>Click to see the details</summary>
@@ -203,11 +217,11 @@ If you want to use WSL2 on Windows 11 you will notice, that Ubuntu WSL2 doesn't
203217

204218
This tutorial will guide you through the process of setting up WSL2 Ubuntu with USB webcam support, rebuilding the kernel, and preparing the environment for the Deep-Live-Cam project.
205219

206-
#### 1. Install WSL2 Ubuntu
220+
**1. Install WSL2 Ubuntu**
207221

208222
Install WSL2 Ubuntu from the Microsoft Store or using PowerShell:
209223

210-
#### 2. Enable USB Support in WSL2
224+
**2. Enable USB Support in WSL2**
211225

212226
1. Install the USB/IP tool for Windows:
213227
[https://learn.microsoft.com/en-us/windows/wsl/connect-usb](https://learn.microsoft.com/en-us/windows/wsl/connect-usb)
@@ -221,7 +235,7 @@ usbipd attach --wsl --busid x-x # Replace x-x with your webcam's bus ID
221235
```
222236
You need to redo the above every time you reboot wsl or re-connect your webcam/usb device.
223237

224-
#### 3. Rebuild WSL2 Ubuntu Kernel with USB and Webcam Modules
238+
**3. Rebuild WSL2 Ubuntu Kernel with USB and Webcam Modules**
225239

226240
Follow these steps to rebuild the kernel:
227241

@@ -234,7 +248,7 @@ Additional info: [https://askubuntu.com/questions/1413377/camera-not-working-in-
234248

235249
3. After rebuilding, restart WSL with the new kernel.
236250

237-
#### 4. Set Up Deep-Live-Cam Project
251+
**4. Set Up Deep-Live-Cam Project**
238252
Within Ubuntu:
239253
1. Clone the repository:
240254

@@ -244,7 +258,7 @@ git clone [https://github.com/hacksider/Deep-Live-Cam](https://github.com/hacksi
244258

245259
2. Follow the installation instructions in the repository, including cuda toolkit 11.8, make 100% sure it's not cuda toolkit 12.x.
246260

247-
#### 5. Verify and Load Kernel Modules
261+
**5. Verify and Load Kernel Modules**
248262

249263
1. Check if USB and webcam modules are built into the kernel:
250264

@@ -271,7 +285,7 @@ dmesg | tail
271285
sudo ls -al /dev/video*
272286
```
273287

274-
#### 6. Set Up Permissions
288+
**6. Set Up Permissions**
275289

276290
1. Add user to video group and set permissions:
277291

@@ -303,15 +317,15 @@ sudo udevadm control --reload-rules && sudo udevadm trigger
303317

304318
5. Start Deep-Live-Cam with `python run.py --execution-provider cuda --max-memory 8` where 8 can be changed to the number of GB VRAM of your GPU has, minus 1-2GB. If you have a RTX3080 with 10GB I suggest adding 8GB. Leave some left for Windows.
305319

306-
#### Final Notes
320+
**Final Notes**
307321

308322
- Steps 6 and 7 may be optional if the modules are built into the kernel and permissions are already set correctly.
309323
- Always ensure you're using compatible versions of CUDA, ONNX, and other dependencies.
310324
- If issues persist, consider checking the Deep-Live-Cam project's specific requirements and troubleshooting steps.
311325

312326
By following these steps, you should have a WSL2 Ubuntu environment with USB webcam support ready for the Deep-Live-Cam project. If you encounter any issues, refer back to the specific error messages and troubleshooting steps provided.
313327

314-
#### Troubleshooting CUDA Issues
328+
**Troubleshooting CUDA Issues**
315329

316330
If you encounter this error:
317331

@@ -340,26 +354,35 @@ sudo apt-get -y install cuda-toolkit-11-8
340354
```
341355
</details>
342356

343-
## Want the Next Update Now?
344-
If you want the latest and greatest build, or want to see some new great features, go to our [experimental branch](https://github.com/hacksider/Deep-Live-Cam/tree/experimental) and experience what the contributors have given.
345357

346-
## TODO
347-
:heavy_check_mark: Support multiple faces feature
358+
## Future Updates & Roadmap
359+
360+
For the latest experimental builds and features, see the [experimental branch](https://github.com/hacksider/Deep-Live-Cam/tree/experimental).
361+
362+
**TODO:**
363+
364+
- [x] Support multiple faces
348365
- [ ] Develop a version for web app/service
349366
- [ ] UI/UX enhancements for desktop app
350367
- [ ] Speed up model loading
351368
- [ ] Speed up real-time face swapping
352369

353-
*Note: This is an open-source project, and we’re working on it in our free time. Therefore, features, replies, bug fixes, etc., might be delayed. We hope you understand. Thanks.*
370+
This is an open-source project developed in our free time. Updates may be delayed.
371+
354372

355373
## Credits
356374

357375
- [ffmpeg](https://ffmpeg.org/): for making video related operations easy
358376
- [deepinsight](https://github.com/deepinsight): for their [insightface](https://github.com/deepinsight/insightface) project which provided a well-made library and models. Please be reminded that the [use of the model is for non-commercial research purposes only](https://github.com/deepinsight/insightface?tab=readme-ov-file#license).
359377
- [havok2-htwo](https://github.com/havok2-htwo) : for sharing the code for webcam
360-
- [GosuDRM](https://github.com/GosuDRM/nsfw-roop) : for uncensoring roop
378+
- [GosuDRM](https://github.com/GosuDRM) : for open version of roop
361379
- [pereiraroland26](https://github.com/pereiraroland26) : Multiple faces support
362380
- [vic4key](https://github.com/vic4key) : For supporting/contributing on this project
363381
- [KRSHH](https://github.com/KRSHH) : For updating the UI
364382
- and [all developers](https://github.com/hacksider/Deep-Live-Cam/graphs/contributors) behind libraries used in this project.
365383
- Foot Note: [This is originally roop-cam, see the full history of the code here.](https://github.com/hacksider/roop-cam) Please be informed that the base author of the code is [s0md3v](https://github.com/s0md3v/roop)
384+
385+
## Thanks to all the contributors
386+
<a href="https://github.com/hacksider/Deep-Live-Cam/graphs/contributors" target="_blank">
387+
<img src="https://contrib.rocks/image?repo=hacksider/Deep-Live-Cam" />
388+
</a>

0 commit comments

Comments
 (0)