A Python library for object detection format conversion
Project description
VisualFlow
VisualFlow is a Python library for object detection that aims to provide a model-agnostic, end to end data solution for your object detection needs. Convert your data, augment it, and inference your models simply with a few function calls!
We have started this library with the vision of providing end to end object detection, from formatting all the way to inferencing multiple types of object detection models.
Our initial version of VisualFlow allows format conversions between PASCAL VOC, COCO and YOLO. Stay tuned for future updates!
Installation
You can install VisualFlow using pip:
pip install visualflow
Usage
Conversions
VisualFlow provides three main conversion functions: to_voc(), to_yolo(), and to_coco(). Here's how you can use them:
Conversion to YOLO Format
To convert from PASCAL VOC or COCO format to YOLO format, use the to_yolo() function.
For VOC to YOLO:
import VisualFlow as vf
vf.to_yolo(in_format='voc',
images='path/to/images',
annotations='path/to/annotations',
out_dir='path/to/output')
For COCO to YOLO:
import VisualFlow as vf
vf.to_yolo(in_format='coco',
images='path/to/images',
out_dir='path/to/output',
json_file='path/to/annotations.json')
Conversion to Pascal VOC Format
To convert from COCO or YOLO format to Pascal VOC format, use the to_voc() function.
For COCO to VOC:
import VisualFlow as vf
vf.to_voc(in_format='coco',
images='path/to/images',
out_dir='path/to/output',
json_file='path/to/annotations.json')
For YOLO to VOC:
import VisualFlow as vf
vf.to_voc(in_format='yolo',
images='path/to/images',
annotations='path/to/annotations',
class_file='path/to/classes.txt',
out_dir='path/to/output')
Conversion to COCO Format
To convert from PASCAL VOC or YOLO format to COCO format, use the to_coco() function.
For VOC to COCO:
import VisualFlow as vf
vf.to_coco(in_format='voc',
images='path/to/images',
annotations='path/to/annotations',
class_file='path/to/classes.txt',
output_file_path='path/to/output.json')
For YOLO to COCO:
import VisualFlow as vf
vf.to_coco(in_format='yolo',
images='path/to/images',
annotations='path/to/annotations',
class_file='path/to/classes.txt',
output_file_path='path/to/output.json')
Make sure to replace 'path/to/images', 'path/to/annotations', 'path/to/classes.txt', and 'path/to/output' with the actual paths to your dataset files and folders.
Augmentations
VisualFlow's powerful data augmentations can enhance your object detection training data. Easily apply these transformations to your dataset with just a few lines of code:
- Cutout: Create up to three random cutouts to encourage robustness and generalization in your models.
- Grayscale: Convert images to grayscale, adding diversity to your training data.
- Brightness: Adjust the brightness of your images, ensuring your models can handle varying lighting conditions.
- Noise: Introduce noise to diversify your dataset and improve model resilience.
- Blur: Apply blurring to images, simulating real-world scenarios and enhancing model adaptability.
- Hue: Adjust the hue of images, enriching color variations and augmenting the dataset.
- Exposure: Manipulate exposure levels to help models cope with different lighting environments.
- Flip90: Perform 90-degree flips for data variation and better model generalization.
- Shear: Apply shear transformations on bounding boxes to augment your dataset and improve model robustness.
- Rotate: Rotate bounding boxes by a specified angle to create diverse training examples.
Some examples are available below
import VisualFlow as vf
vf.cutout(image_dir='path/to/images',
labels_dir='path/to/labels', # optional
output_dir='path/to/output',
max_num_cutouts=3) # optional, set by default
vf.grayscale(image_dir='path/to/images',
labels_dir='path/to/labels', # optional
output_dir='path/to/output')
vf.brightness(image_dir='path/to/images',
labels_dir='path/to/labels', # optional
output_dir='path/to/output',
factor=1.5) # optional, set by default
vf.noise(image_dir='path/to/images',
labels_dir='path/to/labels', #optional
output_dir='path/to/output')
vf.blur(image_dir='path/to/images',
labels_dir='path/to/labels', # optional
output_dir='path/to/output')
vf.hue(image_dir='path/to/images',
labels_dir='path/to/labels', # optional
output_dir='path/to/output')
vf.exposure(image_dir='path/to/images',
labels_dir='path/to/labels', # optional
output_dir='path/to/output',
factor=2.0) # optional, set by default
vf.flip90(image_dir='path/to/images',
labels_dir='path/to/labels',
output_dir='path/to/output')
vf.shear(image_dir='path/to/images',
labels_dir='path/to/labels',
output_dir='path/to/output',
shear_factor= 0.2) # optional, set by default
vf.rotate(image_dir='path/to/images',
labels_dir='path/to/labels',
output_dir='path/to/output', angle=30) # optional, set by default
Inferences
VisualFlow now empowers you to harness the full potential of your YOLO models, making object detection inferencing a seamless part of your workflow. With this new feature, you can confidently evaluate your trained models on your trained models.
Inference with VisualFlow is a breeze. Here's a simple example of how you can perform inferencing on your YOLO models:
import VisualFlow as vf
model_path = "/path/to/your/yolo_model.pt"
inference_dir = "/path/to/your/inference_images"
labels_dir = "/path/to/your/inference_labels"
class_txt = "/path/to/your/class_names.txt"
output_dir = "/path/to/your/output_directory"
vf.yolo_inference(model_path=model_path,
inference_dir=inference_dir,
labels_dir=labels_dir,
class_txt=class_txt,
output_dir=output_dir)
# additional arguments: iou, conf
We understand that each object detection project may require different configurations. Therefore, VisualFlow's inference() function now supports two additional parameters:
- iou: Set to 0.7 by default, this parameter controls the minimum threshold for bounding box overlap.
- conf: Set to 0.5 by default, this parameter determines the minimum confidence level required for an object detection prediction.
UPDATE: Visualflow now supports Facebook's DETR! (Please convert for test dataset into YOLO format first using the conversion functionality above.) You can use it as follows:
import VisualFlow as vf
model_path = "/path/to/your/yolo_model.pt"
inference_dir = "/path/to/your/inference_images"
labels_dir = "/path/to/your/inference_labels"
class_txt = "/path/to/your/class_names.txt"
output_dir = "/path/to/your/output_directory"
vf.detr_inference(model_path=model_path,
inference_dir=inference_dir,
labels_dir=labels_dir,
class_txt=class_txt,
output_dir=output_dir)
Contributing
Contributions are welcome! If you find any issues or have suggestions for improvements, please feel free to open an issue or submit a pull request on GitHub.
License
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file VisualFlow-0.2.4.tar.gz
.
File metadata
- Download URL: VisualFlow-0.2.4.tar.gz
- Upload date:
- Size: 15.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.10.11
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 86c2db6faa1677abcf535b36d8fe6ffc4c940303e77d9b4fd65d6ef3b94f2b42 |
|
MD5 | aeb0cd0dfdf0802fb0114028483eb4b7 |
|
BLAKE2b-256 | cdab4701c1f4173fca9db14b6c84d5f1add642d050527e614b74b44e6f614d36 |
File details
Details for the file VisualFlow-0.2.4-py3-none-any.whl
.
File metadata
- Download URL: VisualFlow-0.2.4-py3-none-any.whl
- Upload date:
- Size: 14.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.10.11
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 6da6129a24c2b47b336a3c0402ec6525d2a6cb8500faf09af16951e3a61addea |
|
MD5 | 7950b4610c6ebd6bba783c3515a880a3 |
|
BLAKE2b-256 | 77752c44d8d92b985cc42d665320545b4b0a91104741486427f687f5eaaaa17f |