-
Notifications
You must be signed in to change notification settings - Fork 22
Description
Hi, thanks for your awesome work! I recently run the public naive SFT and CIL 512 checkpoint following the B.3 Evaluation setting and get the following result, which may be different from paper reported result.
| Method | FVD (2048-16f) | IS (10k) |
|---|---|---|
| pre fs=3 | 434.60 | 13.16 |
| pre fs=8 | 288.01 | 13.31 |
| cli fs=8 | 409.69 | 13.68 |
| sft fs=8 | 373.77 | 13.19 |
I notice the paper said ``we sample 16 frames at 3 fps'', so I change the inference shell script to FS=3 (NOTE that the FS is actually fps in dynamicrafter) but find the pretrained checkpoint have higher fvd than the paper said. So I also add the FS=8 (fps=24/3=8), but I found CIL at the same fps rate has larger fvd than pre.
So I would like to kindly request more detail about the evaluation detail in the paper, does you evaluate fvd on exact 2048 generated videos, or generate more video but only sample 2048 files randomly each time?
Second, I use RAFT to calculate motion score on the generated 2048 videos following the paper description in page 17, I get the following results:
| Method | Motion Score |
|---|---|
| pre fs=3 | 91.5 |
| pre fs=8 | 78.17 |
| cli fs=8 | 73.78 |
| sft fs=8 | 85.89 |
It seems that sft has higher motion score than the pretrained model, which is contrary to the Table1. Could you help me to find problems ?