Skip to main content

A research framework for artificial agent interactions

Project description

InterLab: A research framework for artificial agent interactions

Welcome to InterLab, a research-focused toolkit created to facilitate study and experimentation in the realm of agent interactions, particularly those based on Language Learning Models (LLMs). Our primary objective is to simplify the process of crafting, deploying, and inspecting complex and structured queries within the context of agent interactions, while also providing robust support for interaction logging, UI and visualization. While we maintain a broad scope and plan to include game theoretic agents and a variety of scenarios, our main emphasis lies in the sphere of LLM interactions.

InterLab is developed at the Alignment of Complex Systems Research Group and distributed under MIT license (see LICENSE).

Current status: InterLab is in open beta. While we use InterLab internally for our research, it is also under active development and although we aim to limit any breaking changes to 0.x version changes, the API may change without notice. We plan to announce InterLab with a complete API, use-cases and documentation in Aug 2023, and while we encourage you to use it for your alignment research already, we also ask you not to propagate it widely in the next few weeks.

If you use InterLab, want to share ideas, feedback or have any questions, please email us at gavento@acsresearch.org.

Overview

InterLab is composed of several core modules, each providing distinct functionalities:

  • actor provides framework for actor interactions, including a generic LLM single-shot agent and a web console for playing as an actor, along with actor memory systems. The agents may be queried for any JSON-like action type.
  • environemnts a framework in which actors may interact
  • queries presents powerful helpers for advanced queries to the models: querying the model for structured data for any dataclass or Pydantic model (incl. auto-generating example instances of a schema), and providing robust and comprehensive response parsing for JSON (with repeat and validation options). It also includes simple yet robust summarization.

Interlab is also shipped with independenent module TreeTrace that offers comprehensive structured logging of nested TracingNodess, storage for traing nodes, and custom visualization of content such as Images, generic HTML, and tracking f-string-like field substitutions in larger text (with FormatStr).

Beyond that, you can find more less-tested and experimental code in experimental submodules accross the package.

Structured interaction log browser

In-notebook or independent browser for the structured logs, with live updates, support for JSON-like structured data and inline visualizations. Captures both high-level interaction structure and the low-level API calls for easy inspection and debugging. Example screenshots (click to zoom in):

Tracing browser screenshot

Tracing browser screenshot

Example notebooks

You can find Jupyter notebooks with a few worked-out examples here. The notebook car_negotiation.ipynb is a good starting point for a simple bargaining simulation.

InterLab Zoo

The interlab_zoo package serves as a repository for specific and opinionated implementations of actors, scenarios, actor memory systems, tracing post-processing, and other tools that enhance the InterLab project. Its mission is to gather code that is both widely applicable and useful, while maintaining a compact and focused core package.

Install

Install the package interlab from PyPI using pip or any other package manager.

pip install interlab

Install with development tools

Poetry. This repository utilizes poetry for package management, which is recommended for dependency installation and is required for InterLab development. Poetry automatically generates and manages a virtual environment for you, and also installs interlab module itself. If you have poetry installed, running the following command inside this repository will install InterLab:

poetry install

pip and requirements.txt. Alternatively, pip can be used to install dependencies with pip install -r requirements.txt (core requirements) or pip install -r requirements-full.txt (including development tools, Jupyter Lab, etc.; equivalent to poetry install). This is also required to use InterLab in Google Colab (use requirements-colab.txt to get aroung some Colab versioning conflicts).

Run

Jupyter Lab provides the simplest way to interact with the code and design experiments:

poetry run jupyter lab
# Or without poetry, in the project root folder:
jupyter lab

After running the command, open the provided link in your browser. notebooks/car_negotiation.ipynb is a recommended starting point.

Google Colab

Google Colab often offers a lightweight alternative to setting up InterLab locally on your computer. Interlab comes with built-in colab compatibility and we have prepared a Template InterLab Colab notebook with common setup and a simple example experiment with two LLMs interacting on behalf of their users.

Note: API Keys

In order to use LLM provider serveics and APIs, you need to generate and provide the corresponding API keys. You can provide the keys as environment variables, via .env file, or interactively on every run (e.g. in the colab). Storing keys in the notebook is possible but not recommended (as they easily leak into git or while sharing the code).

API keys can be stored in a .env file located in the notebook directory or your home directory. (This file is ignored by git by default, providing an additional security measure.) The file is a simple text file with key=value pairs, for example:

OPENAI_API_KEY=sk-...
OPENAI_API_ORG=org-...
ANTHROPIC_API_KEY=sk-ant-...
...

You can then import these variables from the .env file into a Jupyter notebook using the dotenv package:

import dotenv
dotenv.load_dotenv()

Roadmap

A rough overview where the project is heading in our minds, incomplete and subject to prioritization based on experimentor's needs. Contact us if you want to get involved in developing interlab with us.

Games, scenarios, game theory

  • Additional general agent implementations
    • Deliberative agent, internal debate agent, constitutional agent, ...
    • Subagent minds implementation
  • Additional memory systems
    • Recall-based memory system
  • Support for counterfactual development and situation checkpointing
    • A stateful scenario (game) system, storing state of games in progress (likely as a pickle)
    • Markov-Chain Monte-Carlo agent strategy (later generalized to regret minimization)
  • Integration with Gymnasium environments (former OpenAI Gym), interactive OpenAI evals, and others (may need mostly wrappers)

Language models

  • Support for retrieving activation vectors from LLMs (e.g. HuggingFace ones), then support for influence vectors
  • Technical: Better support for token counting and limiting tokens of LLMs

More integrations

  • Experiment tracking and management (e.g. Hydra)
  • Visualizations, e.g. Plotly charts
  • Showcase and improve interactive (human player) interfaces

Research LLM sensitivity to framings and formulations in the framework

Listing a few directly related to the framework:

  • How much does asking for JSON change the nature/properties of the responses (vs asking for response in natural language only)?
    • If it tunrs it does, we can do all JSON requests in two steps: first answer in free-form text, then use a second independent request to just extract the response information into the JSON.
    • The request for free-form text reply needs to ask about all the fields of the expected data type. Can this question creation be automated?
  • Other sensitivity to formulations: Use an LLM to generate variants of prompts (GPT-4 can do this) and evaluate differences in outcomes.
  • Evaluate memory recall quality (for vector-embedding memory systems, or other adaptive recall), to tune the systems and be sure information is not lost in longer interactions.

Other ideas and directions

  • Active actor interactions - actors decide when to act instead of being prompted for action, either in simulated or in real time.
  • Network protocol for remote agents, connection to other actor frameworks, games, or UIs.

Contributions and development

To contribute, please submit a pull request with your proposed changes based on the latest main branch.

By submitting a pull request for this project, you agree to license your contribution under the MIT license to this project as written in the LICENSE file.

Citing InterLab

If you use InterLab in your researh, please cite it in your work using the "Cite this repository" github gadget or with:

Tomáš Gavenčiak, Ada Böhm: InterLab [Computer software]. 2023. https://github.com/acsresearch/interlab

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

interlab-0.4.0.tar.gz (732.1 kB view hashes)

Uploaded Source

Built Distribution

interlab-0.4.0-py3-none-any.whl (744.8 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page