Skip to main content

A contextual bandit benchmarking package.

Project description

Coba

An agile, collaborative benchmarking framework for contextual bandit research.

How do you benchmark?

Think for a second about the last time you benchmarked an algorithm or dataset and ask yourself

  1. Was it easy to add new data sets?
  2. Was it easy to add new algorithms?
  3. Was it easy to create, run and share benchmarks?

The Coba Way

Coba was built from the ground up to do all that and more.

Coba is...

  • ... light-weight (it has no dependencies to get started)
  • ... distributed (it was built to work across the web with caching, api-key support, checksums and more)
  • ... verbose (it has customizable, hierarchical logging for meaningful, readable feedback on log running jobs)
  • ... robust (benchmarks write every action to file so they can always be resumed whenever your system crashes)
  • ... just-in-time (no resources are loaded until needed, and they are released immediately to keep memory small)
  • ... a duck? (coba relies only on duck-typing so no inheritance is needed to implement our interfaces)

But don't take our word for it. We encourage you to look at the code yourself, and more documentation will be coming soon.

Adding New Data Sets

Creating Simulations From Classification Data Sets

Classification data sets are an easy way to quickly evaluate CB algorithms. So long as the data set is in CSV format it can easily be turned into a contextual bandit simulation with ClassificationSimulation.from_csv. This method works both with local files and files available over http. When requesting over http Coba even allows you to cache a version of the file locally for fast re-runs later.

Creating Simulations From Generative Models

Certain domains have well defined models that an agent has to make decisions within. To add datasets from these domains one can use LambdaSimulation to define random functions that returns contexts, actions and rewards according to any defined distribution.

Creating Custom Simulations From Scratch

If more customization is needed than what is offered above then a new simulation class can be created. Implementing such a class requires no inheritance or complex registration. One simply has to satisfy Coba's Simulation interface as shown below (with Interaction included for reference):

class Simulation:

    @property
    def interactions(self) -> Sequence[Interaction]:
        """The sequence of interactions in a simulation."""
        ...

    def rewards(self, choices: Sequence[Tuple[Key,int]] ) -> Sequence[float]:
        """The observed rewards for interactions (identified by key) and their selected action indexes."""
        ...

class Interaction:
    @property
    def context(self) -> Hashable:
        """The interaction's context description."""
        ...

    @property
    def actions(self) -> Sequence[Hashable]:
        """The interactions's available actions."""
        ...

    @property
    def key(self) -> Key:
        """A unique key identifying the interaction."""
        ...

Adding New Algorithms

A number of algorithms have been implemented out of the box including epsilon-greedy, VowpalWabbit bagging, VowpWabbit softmax, VowpalWabbit cover, VowpalWabbit RND and upper confidence bounding. Adding algorithms is simply a matter of satisfying the Learner interface as shown below:

class Learner:
    """The interface for Learner implementations."""

    @abstractmethod
    def choose(self, key: int, context: Hashable, actions: Sequence[Hashable]) -> int:
        """Choose which action index to take."""
        ...

    @abstractmethod
    def learn(self, key: int, context: Hashable, action: Hashable, reward: float) -> None:
        """Learn about the result of an action that was taken in a context.
        ...

Creating and Sharing Benchmarks

Benchmarks are created using a json configuration file. In the configuration file one defines the data sets to include in the benchmark, the location of the datasets, how to break the datasets into batches, what random seed to use and if the data sets should be randomized. By placing all these characteristics into a single configuration file creating, modifying and sharing benchmarks is simply a matter of editing this file and emailing it to another researcher.

 {
    "templates"   : { "shuffled_openml_classification": { "seed":777, "type":"classification", "from": {"format":"openml", "id":"$id"} }},
    "batches"     : { "count":51 },
    "ignore_first": true,
    "simulations" : [
        {"template": "shuffled_openml_classification", "$id":3},
        {"template": "shuffled_openml_classification", "$id":6},
        {"template": "shuffled_openml_classification", "$id":8}
    ]
}

Examples

An examples directory is included in the repository with a number of code demonstrations and benchmark demonstrations. These examples show how to create benchmarks, evaluate learners against them and plot the results.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

coba-1.0.3.tar.gz (35.0 kB view hashes)

Uploaded Source

Built Distribution

coba-1.0.3-py3-none-any.whl (41.0 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page