A utility repo for vision dataset access and management.
Project description
Vision Datasets
Introduction
This repo
- defines unified contract for dataset for purposes such as training, visualization, and exploration, via
DatasetManifest
,ImageDataManifest
, etc. - provides many commonly used dataset operation, such as sample dataset by categories, sample few-shot sub-dataset, sample dataset by ratios, train-test split, merge dataset, etc. (See Here)
- provides API for organizing and accessing datasets, via
DatasetHub
Currently, seven basic
types of data are supported:
image_classification_multiclass
: each image can is only with one label.image_classification_multilabel
: each image can is with one or multiple labels (e.g., 'cat', 'animal', 'pet').image_object_detection
: each image is labeled with bounding boxes surrounding the objects of interest.image_text_matching
: each image is associated with a collection of texts describing the image, and whether each text description matches the image or not.image_matting
: each image has a pixel-wise annotation, where each pixel is labeled as 'foreground' or 'background'.image_regression
: each image is labeled with a real-valued numeric regression target.image_caption
: each image is labeled with a few texts describing the images.text_2_image_retrieval
: each image is labeled with a number of text queries describing the image. Optionally, an image is associated with one label.visual_question_answering
: each image is labeled with a number of question-answer pairsvisual_object_grounding
: each image is labeled with a number of question-answer-bboxes triplets.
multitask
type is a composition type, where one set of images has multiple sets of annotations available for different tasks, where each task can be of any basic type.
Note that image_caption
and text_2_image_retrieval
might be merged into image_text_matching
in future.
Dataset Contracts
DatasetManifest
wraps the information about a dataset including labelmap, images (width, height, path to image), and annotations.ImageDataManifest
encapsulates information about each image.ImageDataManifest
encapsulates image-specific information, such as image id, path, labels, and width/height. One thing to note here is that the image path can be- a local path (absolute
c:\images\1.jpg
or relativeimages\1.jpg
) - a local path in a non-compressed zip file (absolute
c:\images.zip@1.jpg
or relativeimages.zip@1.jpg
) or - an url
- a local path (absolute
ImageLabelManifest
: encapsulates one single image-level annotationCategoryManifest
: encapsulates the information about a category, such as its name and super category, if applicableVisionDataset
is an iterable dataset class that consumes the information fromDatasetManifest
.
VisionDataset
is able to load the data from all three kinds of paths. Both 1. and 2. are good for training, as they access data from local disk while the 3rd one is good for data exploration, if you have the data in azure storage.
For multitask
dataset, the labels stored in the ImageDataManifest
is a dict
mapping from task name to that task's labels. The labelmap stored in DatasetManifest
is also a dict
mapping from task name to that task's labels.
Creating DatasetManifest
In addition to loading a serialized DatasetManifest
for instantiation, this repo currently supports two formats of data that can instantiates DatasetManifest
,
using DatasetManifest.create_dataset_manifest(dataset_info, usage, container_sas_or_root_dir)
: COCO
and IRIS
(legacy).
DatasetInfo
as the first arg in the arg list wraps the metainfo about the dataset like the name of the dataset, locations of the images, annotation files, etc. See examples in the sections below
for different data formats.
Once a DatasetManifest
is created, you can create a VisionDataset
for accessing the data in the dataset, especially the image data, for training, visualization, etc:
dataset = VisionDataset(dataset_info, dataset_manifest, coordinates='relative')
Coco format
Here is an example with explanation of what a DatasetInfo
looks like for coco format, when it is serialized into json:
{
"name": "sampled-ms-coco",
"version": 1,
"description": "A sampled ms-coco dataset.",
"type": "object_detection",
"format": "coco", // indicating the annotation data are stored in coco format
"root_folder": "detection/coco2017_20200401", // a root folder for all files listed
"train": {
"index_path": "train.json", // coco json file for training, see next section for example
"files_for_local_usage": [ // associated files including data such as images
"images/train_images.zip"
]
},
"val": {
"index_path": "val.json",
"files_for_local_usage": [
"images/val_images.zip"
]
},
"test": {
"index_path": "test.json",
"files_for_local_usage": [
"images/test_images.zip"
]
}
}
Coco annotation format details w.r.t. image_classification_multiclass/label
, image_object_detection
, image_caption
, image_text_match
and multitask
can be found in COCO_DATA_FORMAT.md
.
Index file can be put into a zip file as well (e.g., annotations.zip@train.json
), no need to add the this zip to "files_for_local_usage" explicitly.
Iris format
Iris format is a legacy format which can be found in IRIS_DATA_FORMAT.md
. Only multiclass/label_classification
, object_detection
and multitask
are supported.
Dataset management and access
Check DATA_PREPARATION.md for complete guide on how to prepare datasets in steps.
Once you have multiple datasets, it is more convenient to have all the DatasetInfo
in one place and instantiate DatasetManifest
or even VisionDataset
by just using the dataset name, usage (
train, val ,test) and version.
This repo offers the class DatasetHub
for this purpose. Once instantiated with a json including the DatasetInfo
for all datasets, you can retrieve a VisionDataset
by
import pathlib
from vision_datasets.common import Usages, DatasetHub
dataset_infos_json_path = 'datasets.json'
dataset_hub = DatasetHub(pathlib.Path(dataset_infos_json_path).read_text(), blob_container_sas, local_dir)
stanford_cars = dataset_hub.create_manifest_dataset('stanford-cars', version=1, usage=Usages.TRAIN)
# note that you can pass multiple datasets.json to DatasetHub, it can combine them all
# example: DatasetHub([ds_json1, ds_json2, ...])
# note that you can specify multiple usages in create_manifest_dataset call
# example dataset_hub.create_manifest_dataset('stanford-cars', version=1, usage=[Usages.TRAIN, Usages.VAL])
for img, targets, sample_idx_str in stanford_cars:
img.show()
img.close()
print(targets)
Note that this hub class works with data saved in both Azure Blob container and on local disk.
If local_dir
:
- is provided, the hub will look for the resources locally and download the data (files included in "
files_for_local_usage", the index files, metadata (if iris format), labelmap (if iris format))
from
blob_container_sas
if not present locally - is NOT provided (i.e.
None
), the hub will create a manifest dataset that directly consumes data from the blob indicated byblob_container_sas
. Note that this does not work, if data are stored in zipped files. You will have to unzip your data in the azure blob. (Index files requires no update, if image paths are for zip files:a.zip@1.jpg
). This kind of azure-based dataset is good for large dataset exploration, but can be slow for training.
When data exists on local disk, blob_container_sas
can be None
.
Operations on manifests {#oom}
There are supported operations on manifests for different data types, such as split, merge, sample, etc. You can run
vision_list_supported_operations -d {DATA_TYPE}
to see the supported operations for a specific data type. You can use the factory classes in vision_datasets.common.factory
to create operations for certain data type.
from vision_datasets.common import DatasetTypes, SplitFactory, SplitConfig
data_manifest = ....
splitter = SplitFactory.create(DatasetTypes.IMAGE_CLASSIFICATION_MULTICLASS, SplitConfig(ratio=0.3))
manifest_1, manifest_2 = splitter.run(data_manifest)
Training with PyTorch
Training with PyTorch is easy. After instantiating a VisionDataset
, simply passing it in vision_datasets.common.dataset.TorchDataset
together with the transform
, then you are good to go with the PyTorch DataLoader for training.
Helpful commands
There are a few commands that come with this repo once installed, such as datset check and download, detection conversion to classification dataset, and so on, check UTIL_COMMANDS.md
for details.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file vision_datasets-1.0.4.tar.gz
.
File metadata
- Download URL: vision_datasets-1.0.4.tar.gz
- Upload date:
- Size: 68.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.9.18
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 902e59023bfa3f85a78e1c8bb7be7fdb9d185a0a93707c21ea1b2797afc8b3f3 |
|
MD5 | 40b95c9ab9b707dd8276e7bbba768776 |
|
BLAKE2b-256 | 2f2e8249ad0623446c6221bce1e423e7c10fef0c85c62ae4a0096ce3a9561e3e |
Provenance
File details
Details for the file vision_datasets-1.0.4-py3-none-any.whl
.
File metadata
- Download URL: vision_datasets-1.0.4-py3-none-any.whl
- Upload date:
- Size: 114.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.9.18
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | d0179fceb4118638d8f57c60a50617e718a1e4ebbd0b42eb63b7c8af2cdd5ca6 |
|
MD5 | f046a4daa7cf71c8183678fc3d30ff48 |
|
BLAKE2b-256 | 8f7c5a87c1e9e3e1ffdac8cabf3cef0a65dd1675c8e825a2543b520b5d9a0650 |