Heuristic distribution of weighted items to bins (either a fixed number of bins or a fixed number of volume per bin). Data may be in form of list, dictionary, list of tuples or csv-file.
Project description
This package contains greedy algorithms to solve two typical bin packing problems, (i) sorting items into a constant number of bins, (ii) sorting items into a low number of bins of constant size. Here’s a usage example
>>> import binpacking
>>>
>>> b = { 'a': 10, 'b': 10, 'c':11, 'd':1, 'e': 2,'f':7 }
>>> bins = binpacking.to_constant_bin_number(b,4) # 4 being the number of bins
>>> print("===== dict\n",b,"\n",bins)
===== dict
{'a': 10, 'b': 10, 'c': 11, 'd': 1, 'e': 2, 'f': 7}
[{'c': 11}, {'b': 10}, {'a': 10}, {'f': 7, 'e': 2, 'd': 1}]
>>>
>>> b = list(b.values())
>>> bins = binpacking.to_constant_volume(b,11) # 11 being the bin volume
>>> print("===== list\n",b,"\n",bins)
===== list
[10, 10, 11, 1, 2, 7]
[[11], [10], [10], [7, 2, 1]]
Consider you have a list of items, each carrying a weight w_i. Typical questions are
How can we distribute the items to a minimum number of bins N of equal volume V?
How can we distribute the items to exactly N bins where each carries items that sum up to approximately equal weight?
Problems like this can easily occur in modern computing. Assume you have to run computations where a lot of files of different sizes have to be loaded into the memory. However, you only have a machine with 8GB of RAM. How should you bind the files such that you have to run your program a minimum amount of times? This is equivalent to solving problem 1.
What about problem 2? Say you have to run a large number of computations. For each of the jobs you know the time it will probably take to finish. However, you only have a CPU with 4 cores. How should you distribute the jobs to the 4 cores such that they will all finish at approximately the same time?
The package provides the command line tool “binpacking” using which one can easily bin pack csv-files containing a column that can be identified with a weight. To see the usage enter
$ binpacking -h
Usage: binpacking [options]
Options:
-h, --help show this help message and exit
-f FILEPATH, --filepath=FILEPATH
path to the csv-file to be bin-packed
-V V_MAX, --volume=V_MAX
maximum volume per bin (constant volume algorithm will
be used)
-N N_BIN, --n-bin=N_BIN
number of bins (constant bin number algorithm will be
used)
-c WEIGHT_COLUMN, --weight-column=WEIGHT_COLUMN
integer (or string) giving the column number (or
column name in header) where the weight is stored
-H, --has-header parse this option if there is a header in the csv-file
-d DELIM, --delimiter=DELIM
delimiter in the csv-file (use "tab" for tabs)
-q QUOTECHAR, --quotechar=QUOTECHAR
quotecharacter in the csv-file
-l LOWER_BOUND, --lower-bound=LOWER_BOUND
weights below this bound will not be considered
-u UPPER_BOUND, --upper-bound=UPPER_BOUND
weights exceeding this bound will not be considered
Install
$ pip install binpacking
Examples
In the repository’s directory
cd examples/
binpacking -f hamlet_word_count.csv -V 2000 -H -c count -l 10 -u 1000
binpacking -f hamlet_word_count.csv -N 4 -H -c count
in Python
import binpacking
b = { 'a': 10, 'b': 10, 'c':11, 'd':1, 'e': 2,'f':7 }
bins = binpacking.to_constant_bin_number(b,4)
print("===== dict\n",b,"\n",bins)
b = list(b.values())
bins = binpacking.to_constant_volume(b,11)
print("===== list\n",b,"\n",bins)
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
File details
Details for the file binpacking-1.5.2.tar.gz
.
File metadata
- Download URL: binpacking-1.5.2.tar.gz
- Upload date:
- Size: 8.7 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.2.0 pkginfo/1.6.1 requests/2.24.0 setuptools/54.2.0 requests-toolbelt/0.9.1 tqdm/4.54.1 CPython/3.8.9
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 9e5254e388670401911c5b6684469ce978964dec12a986bb5d0d33d2b1488d17 |
|
MD5 | f1f1b976981b93fa9d7dd3a5a44622ea |
|
BLAKE2b-256 | c4f8ced0d568105ce48e0453da46e848509ebeb3db1261cdf2ed83ed6e607b82 |