Skip to main content

VerifAI MultiLLM: A module to invoke multiple LLMs concurrently and rank their results

Project description

Multi LLM

VerifAI Implementation of invoking multiple large language models concurrently and ranking results


Quick-Start

For further information look to Running the Multi_LLM Application inside of Usage

Installation

pip3 install multillm

Edit the config file to point to your google and openai API keys. Edit and add your "credentials" of each llm to the config.json file:

"llms": [
                {
                    "file": "bard.py",
                    "class_name": "BARD",
                    "model": "chat-bison@001",
                    * "credentials": "/path/to/google/key.json" *
                },

Edit the config file to add or change RAG parameters under VectorStore Key.

"VectorStore": {
            "chunk_size": 2000,
            "chunk_overlap": 100,
            "max_chunk_size": 6144,
            "db_path": "./ChromaDB",
            "k_matches": 1,
            "credentials_openai": "/path/to/openai/key.json"
        }
Example google-app-credentials.json
 {
  "client_id": "123489-6qr4p.apps.googleusercontent.com",
  "client_secret": "fx-d3456-tryf0g9f9",
  "quota_project_id": "my-llm-training",
  "refresh_token": "1-34GFH89KLwe-eft",
  "type": "authorized_user"
  }
Example openai-credentials.json
 {
    "organization" : "org-jc8901FDLI0267",
    "api_key" : "rt-067FGDiTL834"
}
  

Basic Example

multillm -c config.json -prompt "write a python function to find a root of the function f using Newton's method"

The above example will run the config.json with the provided prompt.

Output of the above multillm command
loading module bard...
finished loading module bard
registered model BARD <bard.BARD object at 0x10e4e5b70>
loading module GPT...
finished loading module GPT
registered model GPT <GPT.GPT object at 0x10e4e5d20>
loaded llms: {'bard': <module 'bard' from '/Users/Verifai/models/bard.py'>, 'GPT': <module 'GPT' from '/Users/Verifai/models/GPT.py'>}
calling model: BARD
calling model: GPT
model chat-bison@001

BARD Response:

def newton(f, df, x0, tol=1e-6, maxiter=100):
    """
    Find a root of the function f using Newton's method.

    Args:
        f: The function to find a root of.
        df: The derivative of f.
        x0: The initial guess for the root.
        tol: The tolerance for convergence.
        maxiter: The maximum number of iterations to perform.
        

    Returns:
        The root of f, or None if no root was found.
    """

    for i in range(maxiter):
        x1 = x0 - f(x0) / df(x0)
        if abs(x1 - x0) < tol:
            return x1
        x0 = x1
    return None

GPT Response:

def newton_method(f, f_prime, initial_guess, tol=1e-6, max_iter=100):
    """
    Newton's method for finding the root of a function.

    Parameters:
        f (function): The function for which the root is to be found.
        f_prime (function): The derivative of f.
        initial_guess (float): The initial guess for the root.
        tol (float): The desired tolerance (default 1e-6).
        max_iter (int): The maximum number of iterations (default 100).

    Returns:
        float: The root found by Newton's method, or None if no root is found.
    """
    x = initial_guess
    for _ in range(max_iter):
        fx = f(x)
        if abs(fx) < tol:
            return x
        fpx = f_prime(x)
        if fpx == 0:
            return None
        x -= fx / fpx
    return None

MultiLLM Explanation and Ranking for :

  • "The GPT implementation of the Newton's method function (newton_method) is well-documented with clear parameter explanations and return format. It also includes appropriate input validation checks and handles the case when the derivative is 0. Overall, it is a comprehensive and robust implementation."

  • "The BARD implementation of the Newton's method function (newton) is also well-documented and accepts the necessary parameters. However, it lacks input validation checks, such as when the derivative is 0, and does not provide an explicit return value when no root is found. It can be improved by addressing these limitations."

"Based on these factors, I would rank GPT higher than BARD in terms of the clarity, completeness, and robustness of the implementation."


Development

Please follow below instructions to install and run the repo locally for development and testing:

# Clone the repo to local machine
git clone https://github.com/verifai/mllm.git

cd <your-repo-directory>

# Create a new python virtual environment
python3 -m venv myenv
source myenv/bin/activate

# install the dependencies
pip install -r multillm/requirements.txt

# prompt the multillm with the repo (not the package / multillm command) in macOS:
python run_multillm.py -c config.json -prompt "How to learn python? Answer me within 100 words." -debug -macos

# prompt the multillm with the repo (not the package / multillm command) in Linux:
python run_multillm.py -c config.json -prompt "How to learn python? Answer me within 100 words." -debug

# start the flask server (But for running the command, we don't need to start the flask server)
python serve.py

Usage

The Multi_LLM application provides a powerful and efficient solution for invoking multiple large language models (LLMs) concurrently and effectively managing their outputs. This section guides you through the steps required to run the application using various command-line options and configuration files.

Prerequisites

Before running the Multi_LLM application, ensure that you have the following prerequisites in place:

  • Python >= 3.8 installed on your system.

Command-Line Usage

The Multi_LLM application can be executed from the command line using the following command structure:

multillm -c <config_file> -prompt "<prompt_text>"
  • <config_file>: The path to the configuration file containing LLM details.
  • <prompt_text>: The prompt text you want to provide to the language models.

Config Files

The configuration file config.json defines the language models and their associated parameters. See Config JSON for more information. It follows the structure outlined below:

Example Config.json file
{
    "Config": {
        "Multi_LLM": {
            "llms": [
                {
                    "file": "bard.py",
                    "class_name": "BARD",
                    "model": "chat-bison@001",
                    "credentials": "/path/to/google/key.json"
                },
                {
                    "file": "GPT.py",
                    "class_name": "GPT",
                    "model": "gpt-3.5-turbo",
                    "credentials": "/path/to/openai/key.json"
                }
            ]
        },
        "VectorStore": {
            "chunk_size": 2000,
            "chunk_overlap": 100,
            "max_chunk_size": 6144,
            "db_path": "./ChromaDB",
            "k_matches": 1,
            "credentials_openai": "/path/to/openai/key.json"
        }
    }
}

Running the Application

To run the Multi_LLM application, follow these steps:

  1. pip install multillm

  2. Download the config.json file with the desired language model configurations.

  3. Execute the following command, replacing <config_file> with the actual path to your configuration file and <prompt_text> with the desired prompt text:

    multillm -c <config_file> -prompt "<prompt_text>"
    
  4. The application will run the specified language models concurrently, process their responses using the provided prompt, and display the results.


Adding a new LLM interface

This section will guide you through the process of adding a new LLM by extending the BaseLLM class and customizing it to fit the requirements of your specific language model.

Follow these steps to add an new LLM interface

  1. Start by creating a new Python file in your project directory '<NewLLM.py>' , or within the appropriate package, where you'll define a new class that inherits from BaseLLM. Implement the required methods: get_response() and get_content(). The get_response() method should execute your language model with the provided prompt, and the get_content() method should extract relevant content from the response.

    • See Example Below:
    Example NewLLM.py
    import os,sys
    from multillm.BaseLLM import BaseLLM
    from multillm.Prompt import Prompt
    # <add additional imports here>
    
    
    # NewLLM interface                                                                                                                              
    """                                                                                                                                                 
    The NewLLM class extends the BaseModel class and overrides the get_response() method, providing an implementation.                                           
    """
    class NewLLM(BaseLLM):
        # ... (attributes and __init__ method)
        def __init__ (self, **kwargs):
    	 # add values here directly or if kwargs are specified they are taken from the config file
            defaults  = {
                "class_name" : "NewLLM",
                "model" : "newLLM-bison@06",
                "credentials" : "/path/to-my/key.json"
            }
        # ... Call API and get response from NewLLM
        def get_response(self, prompt):
            # Implement your language model interaction here
            # access credentials file from *self.credentials*
            # access model from *self.model*
      	# access class_name from *self.class_name*
            response =  <"Generated response from  NewLLM model based on prompt">
            return response
    
        # ... Parse and Filter raw response from NewLLM and return text/code content
        def get_content(self, response):
            # Implement content extraction from the response
            content = "Extracted content from response"
            return content
    
  2. Add NewLLM in config.json file, in the 'llms' section

    Add NewLLM to the config.json file
    {
        "Config": {
            "Multi_LLM": {
                "llms": [
                    {
                        "file": "/full-path/NewLLM.py",
                        "class_name": "NEWLLM",
                        "model": "chat-bison@001",
                        "credentials": "/path/to/google/key.json"
                    }, .... ]
    	}
    }
    
  3. Call multillm to run your new LLM or embedd it in your code:

    • multillm -c config.json -prompt "wite a function to sort a billion integers"
  4. (Optional) Embedd NewLLM.py in your code

    • You can now use your custom NewLLM class in your application code. Instantiate it, call its methods, and integrate it into your application's workflow.
custom_llm = NewLLM(model="custom_model", credentials="your_credentials")
prompt = "Generate something amazing."
response = custom_llm.get_response(prompt)
content = custom_llm.get_content(response)
print(content)

By extending the provided BaseLLM class, you can easily create custom language model implementations tailored to your project's needs. This structured approach ensures consistency and modularity in your codebase, allowing you to focus on the unique aspects of your language model while leveraging the foundational structure provided by BaseLLM.


Example of the GPT interface (included with multillm release)

Example Model GPT.py
	
import os,sys
import openai
import json
from multillm.BaseLLM import BaseLLM
from multillm.Prompt import Prompt


# Openai gpt interface
"""
The GPT class extends the BaseModel class,  implements the required methods: get_response() and get_content().
The get_response() method takes a response parameter and returns the content of the first response in the given response object.
"""
class GPT(BaseLLM):

    #implement here
    def __init__ (self, **kwargs):
        # add values here directly or if kwargs are specified they are taken from the config file
        defaults  = {
            "class_name" : "GPT",
            "model" : "gpt-3.5-turbo",
            "credentials" : "key.json"
        }

    # Get Content -- Required Method
    def get_content(self, response):
    
        """ Get the text from the response of an LLM
        e.g.: openai returns the following response, this method should return the 'content'.
        { "choices": [{{"message": 
			{"content": "def binary_sort(arr):"}}}]}
        """
        return response["choices"][0]["message"]["content"]

    # Get Response -- Required Method, Call openai API
    def get_response(self, prompt):
        # setup prompt for API call
        messages=[]
        
        messages.append( {"role": prompt.get_role(), "content" : prompt.get_string()})
        if prompt.context:
            messages.append({"role": prompt.get_role(), "content" : prompt.get_context()})
        
        # Read Credentials file specified in the config.json, setup for openai
        if not os.path.exists(self.credentials):
            print('error (multi_llm): could not find openai_credentials: {0}' .format(self.credentials))
            return 

        # Open the file for reading
        try:
            with open(self.credentials, 'r') as file:
                # Load the JSON data from the file
                data = json.load(file)
                openai.organization = data['organization']
                openai.api_key = data['api_key'] 

        except Exception as e:
            print('(multi_llm) error: could not load credentials {0} : {1}' .format(self.credentials,str(e)))
            return
                    
        # do API call
        response = openai.ChatCompletion.create(
            model = self.model,
            messages=messages
        )
        if response:
            return(self.get_content(response))
        else:
            return response

Config JSON

Using the config.json File

The config.json file offers a convenient way to configure and load multiple language models using the "Multi_LLM" framework. This section will guide you through the process of creating and utilizing a config.json file to load and use specific language models in your application.

Configuration Setup

To get started, follow these steps to configure your config.json file:

  1. Create a Configuration File: Create a new file named config.json in your project directory or the desired location.

  2. Configure LLMs: Define the language models you want to use within the "llms" array. Each model configuration includes details such as the Python file containing the model implementation, class name, model name, and credentials file path.

{
    "Config": {
        "Multi_LLM": {
            "llms": [
                {
                    "file": "bard.py",
                    "class_name": "BARD",
                    "model": "chat-bison@001",
                    "credentials": "/path/to/google/key.json"
                },
                {
                    "file": "GPT.py",
                    "class_name": "GPT",
                    "model": "gpt-3.5-turbo",
                    "credentials": "/path/to/openai/key.json"
                }
            ]
        }
    }
}

Loading Models Using MultiLLM

After creating the config.json file, you can use the "Multi_LLM" framework to load and utilize the configured models in your application. Follow these steps:

  1. Instantiate Multi_LLM: Create an instance of the Multi_LLM class and provide the path to your config.json file.
from Multi_LLM import Multi_LLM

# Specify the path to your config.json file
config_path = "path/to/config.json"

# Instantiate Multi_LLM
multi_llm = Multi_LLM(config=config_path)
  1. Run Models: Use the run method to run the loaded models and process their responses. Provide a prompt and an optional action chain if needed. The responses from each model will be returned in a dictionary.
prompt = "Translate this sentence."
action_chain = None  # You can define an action chain if required

responses = multi_llm.run(prompt, action_chain)
print(responses)

Action Class

This is the interface class which we use to operate on the output while still in parallel. Action class instances are define by the user and can be chained indefinitely with other Actiion instances. There are two methods that are used to interface with the class: apply() and then(). The first, apply() is what is used internally to call the methods we register. Registration is as follows:

# Write an interfacing function, a simple in/out
def capitalize(data):
	data = data.upper()
	return data
	
def lower(data):
	data = data.lower()
	return data
	
# Create the respective objects
action1 = Action(operation=capitalize)
action2 = Action(operation=lower)

# Create the chain of actions
action_chain = action1.then(action2)

What we did was create some actions that are simple I/O operations on data. We can then chain them together using the then() method mentioned above. The order of operations for then is left to right, in this case action1 will go first then action2. These actions can be tailored to your own specification, since the actions are serial, you can specify the information that is going to be passed into the function and what information will be returned.

After you have created the action chain you can pass this into the MultiLLM.run() method and run.

Action Chain

The Action Chain is a core component of processesing LLM output. In the provided example.py we see the first action defined is for processing the LLM response and extracting the code alone from the response. This could be a first of many steps, what could follow could be saving the code to the file or extracting information through the ast module. The actions are meant to be ran on each of the model's outputs so they should be generalized for use.

Rank Class

The Rank class is identical in functionality to the Action class, though the use is different. While the Action class is used to accept and modify data in each of the models, the Rank class' purpose is to modify the final combined output of the LLMs. This output is stored in a dictionary keyed by the model's names. The Rank class's methods should be written to take this data and operate on it.

#Write an interfacing function, a simple print
def print_llms(dictionary):
	for key,val in dictionary.items():
		print(key, val)

# Create Rank instances
rank_object = Rank(operation=print_llms)

# Running the LLM, assuming an instance of 
# MultiLLM named mLLM
results = mLLLM.run(prompt, action_chain, rank_object)

In the above code we are doing very similar actions as those seen above. The methodoly is the same. In this case we only have a single object, so instead of creating a rank_chain we simply pass in our single instantiated object.

Architecture

High Level Overview

alt

MultiLLM Class

This is the highest level, here we can instantiate MultiLLM objects using either config files (see config files) or manually instantiating a custom or hosted LLM. In this class we call multiple LLMs concurrently and then we can operate on the results of each in parallel using the Action Class.

The "MultiLLM" Python code provides a versatile framework for managing and orchestrating multiple language models (LLMs) within a single application. This code is designed to enhance the efficiency of working with language models by enabling concurrent execution, response processing, and model loading from configuration files. The key features and components of the code include:

  • Concurrent Model Execution: The code allows you to run multiple language models concurrently, facilitating efficient utilization of computational resources. This is particularly useful for scenarios that involve processing multiple prompts or tasks simultaneously.

  • Action Chain Processing: The framework supports the concept of "action chains," which enable you to preprocess model responses using a sequence of predefined actions. This empowers you to refine and enhance the output generated by the language models.

  • Model Loading from Configuration: You can load LLMs from a configuration file. This JSON-based configuration includes essential details about each model, such as the model class, associated credentials, and file paths. The code can dynamically load these models, making it easy to add new models or modify existing ones without altering the core codebase.

  • Redis Integration: The code features optional integration with a Redis instance. It checks for a successful Redis connection and adjusts its behavior accordingly. If Redis connectivity is not established, the code gracefully handles the situation by setting a flag that reflects the absence of a Redis connection.

  • Simplified Model Registration: The code includes a straightforward method to register models within the framework. This allows for the inclusion of custom LLM implementations while ensuring that the models are appropriately organized and accessible for execution.

By using the "MultiLLM" Python code, developers can streamline their interactions with multiple language models, seamlessly integrating them into various applications or projects. The code promotes modularity, reusability, and parallelism in working with language models, ultimately enhancing the user experience and productivity.

BaseLLM

BaseLLM is designed to serve as the basis for implementing various language model classes. This BaseLLM class encapsulates essential attributes and methods necessary for interfacing with language models. The code establishes a structured foundation for building specific language model implementations and harmonizes their interaction within a larger application context.

Key features and components of the code include:

  • Attributes for Language Models: The BaseLLM class declares a set of attributes, such as model, roles, messages, temp, api_key, max_tokens, and args, that are pertinent to language models. These attributes are meant to be customized and adapted as needed for specific model implementations.

  • Customizable Initialization: The __init__() method in the BaseLLM class facilitates flexible instantiation of model instances by allowing the specification of custom values via keyword arguments. It enables the convenient configuration of attributes like name, credentials, model, and class_name.

  • Placeholder Methods: The BaseLLM class defines two placeholder methods: get_response() and get_content(). The get_response() method is designed to receive a Prompt object and run the associated language model with the provided prompt. The actual implementation of this method is expected to be customized in derived classes to perform the model-specific interactions. Similarly, the get_content() method is meant to be implemented by deriving classes, providing an interface to extract relevant content from model responses.

  • Structured Framework: The code encapsulates a structured framework that abstracts common functionalities of language models. By inheriting from the BaseLLM class, developers can focus on implementing model-specific interactions while leveraging the established structure for attribute handling and method placeholders.

BaseLLM aims to streamline the development of specific language model implementations by providing a consistent structure and standardized attributes. Developers can extend this base class to create custom language model classes that seamlessly integrate into the broader application ecosystem. This modular approach promotes reusability, maintainability, and consistent design patterns when working with various language models.


Contribution

You can contribute by extending the models located in models. See the BaseLLM section for more information on the necessary mechanisms for extending the BaseLLM class. For further information contact ethansaurusrex

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

multillm-0.1024.tar.gz (44.9 kB view details)

Uploaded Source

Built Distribution

multillm-0.1024-py3-none-any.whl (61.4 kB view details)

Uploaded Python 3

File details

Details for the file multillm-0.1024.tar.gz.

File metadata

  • Download URL: multillm-0.1024.tar.gz
  • Upload date:
  • Size: 44.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.8.9

File hashes

Hashes for multillm-0.1024.tar.gz
Algorithm Hash digest
SHA256 0034bed944ae20a6e5f6cf287f1daf19003a389044635c7e2e35841a67c2d6b1
MD5 8003ca775c0039bccba25f83a2bf0752
BLAKE2b-256 1e394c5a4924d3d86808f3bd23f1642311f407329ed6dcf1bf1329a0d93aeb9a

See more details on using hashes here.

File details

Details for the file multillm-0.1024-py3-none-any.whl.

File metadata

  • Download URL: multillm-0.1024-py3-none-any.whl
  • Upload date:
  • Size: 61.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.8.9

File hashes

Hashes for multillm-0.1024-py3-none-any.whl
Algorithm Hash digest
SHA256 a34f45483bdd4d1f2e202c792da70a2576d333e3c2e8ba2f6de91c45ed36784c
MD5 16e7c5018ddc1ee2e992286c652ac31d
BLAKE2b-256 41e23c626cd1d78223ecd9314675e46923e18d83324101a6065d71c0c3c5b519

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page