All-in-one encode toolkit
Project description
Av1an
A cross-platform all-in-one tool for streamlining AV1 encoding
Easy, Fast, Efficient and Feature Rich
An easy way to start using AV1 / VP9 / VP8 encoding. AOM, rav1e, SVT-AV1, VPX are supported.
Example with default parameters:
av1an -i input
With your own parameters:
av1an -i input -enc aom -v " --cpu-used=3 --end-usage=q --cq-level=30 --threads=8 " -w 10 -p 2
-ff " -vf scale=-1:720 " -a " -c:a libopus -b:a 24k " -s scenes.csv -log my_log -o output
Usage
-i --file_path Input file(s) (relative or absolute path). Will be processed with same
settings.
-o --output_file Name/Path for output file (Default: (input file name)_av1.mkv)
Output file ending is always `.mkv`
-enc --encoder Encoder to use (aom,rav1e,svt_av1,vpx. Default: aom)
Example: -enc rav1e
-v --video_params Encoder settings flags (If not set, will be used default parameters.
Required for SVT-AV1s)
Must be inside ' ' or " "
-p --passes Set number of passes for encoding
(Default: AOMENC: 2, rav1e: 1, SVT-AV1: 2, VPX: 2)
At current moment 2nd pass rav1e not working
-a --audio_params FFmpeg audio settings flags (Default: copy audio from source to output)
Example: -a '-c:a libopus -b:a 64k'
-w --workers Overrides automatically set number of workers.
Example: rav1e settings " ... --tile-rows 2 --tile-cols 2 ... " -w 3
--split_method Method used for generating splits.(Default: PySceneDetect)
Options: `pyscene`, `aom_keyframes`
`pyscene` - PyScenedetect, content based scenedetection
with threshold.
`aom_keyframes` - using stat file of 1 pass of aomenc encode
to get exact place where encoder will place new keyframes.
-tr --threshold PySceneDetect threshold for scene detection Default: 50
-s --scenes Path to file with scenes timestamps.
If given `0` spliting will be ignored
If file not exist, new will be generated in current folder
First run to generate stamps, all next reuse it.
Example: "-s scenes.csv"
-xs --extra_split Adding extra splits if frame distance beetween splits bigger than
given value. Split only on keyframes. Works with/without PySceneDetect
Example: 1000 frames video with single scene,
-xs 200 will try to add splits at keyframes
that closest to 200,400,600,800.
-cfg Save/Read config file with encoder, encoder parameters,
FFmpeg and audio settings.
-ff --ffmpeg FFmpeg options. Applied to each segment individually.
Example:
--ff " -r 24 -vf scale=320:240 "
-fmt --pix_format Setting custom pixel/bit format(Default: 'yuv420p')
Example for 10 bit: 'yuv420p10le'
Encoding options should be adjusted accordingly.
--resume If encode was stopped/quit resumes encode with saving all progress
Resuming automatically skips scenedetection, audio encoding/copy,
spliting, so resuming only possible after actuall encoding is started.
/.temp folder must be presented for resume.
--no_check Skip checking numbers of frames for source and encoded chunks.
Needed if framerate changes to avoid console spam.
By default any differences in frames of encoded files will be reported.
--keep Not deleting temprally folders after encode finished.
-log --logging Path to .log file(Default: no logging)
--temp Set path for temporally folders. Default: .temp
--boost Enable experimental CQ boosting for dark scenes. Refer to 1.7 release notes.
-bl CQ limit for boosting. CQ can't get lower than this value.
-br CQ range for boosting. Delta for which CQ can be changed
--vmaf Calculate vmaf for each encoded clip.
Saves plot after encode, showing vmaf values for all frames,
mean, 1,25,75 percentile.
Requires: Installed FFMPEG and installed libvmaf.
--vmaf_path Custom path to libvmaf models, by default used system one.
--vmaf_target Vmaf value to target. Currently aomenc only. Best works with 85-97.
When using this mode specify full encoding options.
` --end-usage=q ` or ` --end-usage=cq `
with ` --cq-level=N ` must be specified.
--vmaf_steps Number of evenly spaced probes that is used to interpolate vmaf to cq change.
Must be bigger than 3. Optimal is 4-6 probes. Default: 4
--min_cq, --max_cq Minimum and maximum CQ values used in interpolation in target Vmaf mode
Use to limit CQ values range. Default: 25, 50.
Main Features
Spliting video by scenes for parallel encoding because AV1 encoders currently not good at multithreading, encoding is limited to single or couple of threads at the same time.
- PySceneDetect used for spliting video by scenes and running multiple encoders.
- Fastest way to encode AV1 without losing quality, as fast as many cores cpu have :).
- Target VMAF mode. Targeting end result visual quality.
- Resuming encoding without loss of encoded progress.
- Simple and clean console look.
- Automatic determination of how many workers the host can handle.
- Building encoding queue with bigger files first, minimizing waiting for last scene to encode.
- Both video and audio transcoding with FFmpeg.
- Logging of progress of all encoders.
- "Boosting" quality of dark scenes based on their brightness.
Install
-
Prerequisites:
- Install Python3
For Windows in the installer check the option toadd Python to PATH
- Install FFmpeg
- Install AOMENC
- Install rav1e
- Install SVT-AV1
- Install vpx VP9, VP8 encoding
- Install Python3
-
With a package manager:
-
Manually:
- Clone Repo or Download from Releases
python setup.py install
-
Also: On ubuntu systems packages
python3-opencv
andlibsm6
are required
I have no money
Bitcoin - 1gU9aQ2qqoQPuvop2jqC68JKZh5cyCivG
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Av1an-1.10.3.tar.gz
(20.7 kB
view hashes)
Built Distribution
Av1an-1.10.3-py3-none-any.whl
(24.5 kB
view hashes)