A high-level Deep Learning framework that extends PyTorch and PyCUDA.
Project description
ice-learn
ice
is a sweet extension of PyTorch, a modular high-level deep learning framework that extends and integrates PyTorch and PyCUDA with intuitive interfaces. We aims not only to minimize the boilerplate code without loss of functionality, but also maximize the flexibility and usability for extending and composing any deep learning tasks into an integrate multi-task learning program.
NOTE: It is currently in pre-alpha versions, and the API is subject to change.
Features
-
Minimize Boilerplates: You don't need to repeat yourself.
- Config Once, Use Everywhere: Every mutable class can be converted into a
configurable
. Configuration for deep learning project has never been this easy before. A tagging system to manage and reuse any type of resources you need. - Inplace Argument Parser: You can parse command line argument instantly without a long page of previous definition.
- Config Once, Use Everywhere: Every mutable class can be converted into a
-
Maximize Flexiblility: Painless and Incremental Extension from CUDA to non-standard data-preprocessing and training schedules for multi-task learning.
- The kernel data structure of
ice
is a Hypergraph that manages different module nodes (e.g.ice.DatasetNode
,ice.ModuleNode
, etc.) that are switchable between multiple user-defined execution modes. Extending a new dataset, network module or loss function is by adding newnn.Dataset
s,nn.Module
s and pythoncallable
s to specific mode of the entire graph. - We provide PyCUDA support by automatically managing the PyCUDA context as well as providing a simplified
torch.Tensor
class wrapper that supports efficient multi-dimensional element access in CUDA codes. This feature manages to make writing, compile, execution and testing CUDA extensions for PyTorch extremely fast. We also provide a VSCode extension for PyCUDA docstring highlight. - We support Multi-Task Learning training by finding the Pareto Optimal for each task weight so that you do not need to tune them manually. (TODO)
- We support dill-backended Elastic Multiprocessing launch and management that is compitable with Lambda Function and Closures. You can not only build multi-gpu or multi-machine Data Distributed Parallel training program without effort, but also doesn't require to concern about pickability of any part of program in your application. We actually suggest heavy use of lambda functions such as for simple input and output transforms of modules. This feature also contributes to the minimal boilerplates aim of
ice
.
- The kernel data structure of
Install
pip install ice-learn
(Recommended)
or pip install .[dev]
after a git-clone for developers.
Documentation
You can access documentation through Online Documentation Site, or the docs
subdirectory directly. The documentation is partial auto-generated from comment, and partial manually written, the note on how we produce the documenation can be found here.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file ice-learn-0.4.7.tar.gz
.
File metadata
- Download URL: ice-learn-0.4.7.tar.gz
- Upload date:
- Size: 90.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.0 CPython/3.9.12
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 72f431f863e5273ea5b28d56d5dd96b853cb029e155354f6d31f9d3a9cee7d5b |
|
MD5 | 05aa08c5a2442680a5e34928e815cfe6 |
|
BLAKE2b-256 | fac96dc68d4787f2914e33ac1f94e4475d0755b5f47d42c76a3d6098f41a7e29 |
File details
Details for the file ice_learn-0.4.7-py3-none-any.whl
.
File metadata
- Download URL: ice_learn-0.4.7-py3-none-any.whl
- Upload date:
- Size: 109.8 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.0 CPython/3.9.12
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 69620d943ef576c91829871b202a7c1687d990dfa999e17cd397d099387bcfdc |
|
MD5 | 867d2febbd8be4bba5c4228f5ffef930 |
|
BLAKE2b-256 | 2fa037957fbb20a1411ca2b133850131b080008fd8cbd11f9d7fae963012c78c |