DistilDRBA function for VapourSynth
Project description
VS-DistilDRBA
vapoursynth version for DistilDRBA based on vs-rife.
This project is modified from HolyWu/vs-rife and achieves nearly the same interpolation quality as the original DistilDRBA project.
With TensorRT integration, it achieves a 400% acceleration, enabling real-time playback on high-performance NVIDIA GPUs.
Dependencies
- PyTorch 2.6.0 or later
- VapourSynth R66 or later
trt requires additional packages:
- TensorRT 10.7.0.post1 or later
- Torch-TensorRT 2.6.0 or later
To install the latest stable version of PyTorch, Torch-TensorRT and cupy, run:
pip install -U packaging setuptools wheel
pip install -U torch torchvision torch_tensorrt --index-url https://download.pytorch.org/whl/cu128 --extra-index-url https://pypi.nvidia.com
Installation
pip install -U vsdrba_distilled==1.0.0
Note: Please make sure to install all dependencies listed in the Dependencies section before performing the steps in the Installation section.
If you want to download all models at once, run python -m vsdrba_distilled. If you prefer to only download the model you
specified at first run, set auto_download=True in drba_distilled().
Usage
from vsdrba_distilled import drba_distilled
ret = drba_distilled(clip, trt=True, factor_num=2, factor_den=1, scale=1.0, model="v1", auto_download=True)
See __init__.py for the description of the parameters.
Benchmarks
| model | scale | os | hardware | arch | speed(fps) 720 | speed(fps) 1080 | vram 720 | vram 1080 | backend | verified output | batch | level | streams | threads | trtexec shape | precision | usage |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| drba_distilled v1 | 2x | Linux | rtx5070 / 14600kf | drba_distilled | 251 | 115 | 1.8gb | 2.9gb | torch+trt cu128 | yes, works | 1 | 5 | - | 1 | static | RGBH | drba_distilled(clip, trt=True, model="v1", trt_optimization_level=5) |
| drba_distilled v2_lite | 2x | Linux | rtx5070 / 14600kf | drba_distilled | 999+ | 700 | - | - | torch+trt cu128 | yes, works | 1 | 5 | - | 1 | static | RGBH | drba_distilled(clip, trt=True, model="v1", trt_optimization_level=5) |
🤗 Acknowledgement
This project is supported by SVFI Development Team.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file vsdrba_distilled-1.0.0.tar.gz.
File metadata
- Download URL: vsdrba_distilled-1.0.0.tar.gz
- Upload date:
- Size: 16.8 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.9.13
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
6e5d9041dac1c4532c6a329dd9ab874df659e7487fecf9fdb6a7490c14a1a590
|
|
| MD5 |
6c4119907e5c9071ddf9fa10da0ce761
|
|
| BLAKE2b-256 |
328dc31abc37c85f11ced6256ed702a3cd7466490c877491fd8937d50c3a1a3b
|
File details
Details for the file vsdrba_distilled-1.0.0-py3-none-any.whl.
File metadata
- Download URL: vsdrba_distilled-1.0.0-py3-none-any.whl
- Upload date:
- Size: 18.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.9.13
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
07a8f1875eb3a8842f7dc433954e80338df6d1251aa429cad737ffe84b1d21c9
|
|
| MD5 |
334a6816ada9b2290cf0f14c695c6d71
|
|
| BLAKE2b-256 |
544a24d78990e980434e514c2caf24e1525137c43534d015d232d325f842db78
|