Skip to main content

llama-index llms monsterapi integration

Project description

LlamaIndex Llms Integration: Monsterapi

MonsterAPI LLM.

Monster Deploy enables you to host any vLLM supported large language model (LLM) like Tinyllama, Mixtral, Phi-2 etc as a rest API endpoint on MonsterAPI's cost optimised GPU cloud.

With MonsterAPI's integration in Llama index, you can use your deployed LLM API endpoints to create RAG system or RAG bot for use cases such as: - Answering questions on your documents - Improving the content of your documents - Finding context of importance in your documents

Once deployment is launched use the base_url and api_auth_token once deployment is live and use them below.

Note: When using LLama index to access Monster Deploy LLMs, you need to create a prompt with required template and send compiled prompt as input.

See LLama Index Prompt Template Usage example section for more details.

see (https://developer.monsterapi.ai/docs/monster-deploy-beta) for more details

Once deployment is launched use the base_url and api_auth_token once deployment is live and use them below.

Note: When using LLama index to access Monster Deploy LLMs, you need to create a prompt with reqhired template and send compiled prompt as input. see section LLama Index Prompt Template Usage example for more details.

Examples:

pip install llama-index-llms-monsterapi

  1. MonsterAPI Private LLM Deployment use case

    from llama_index.llms.monsterapi import MonsterLLM
    
        llm = MonsterLLM(
            model = "<Replace with basemodel used to deploy>",
            api_base="https://ecc7deb6-26e0-419b-a7f2-0deb934af29a.monsterapi.ai",
            api_key="a0f8a6ba-c32f-4407-af0c-169f1915490c",
            temperature=0.75,
        )
    
        response = llm.complete("What is the capital of France?")
        ```
    
  2. Monster API General Available LLMs

    from llama_index.llms.monsterapi import MonsterLLM
    
        llm = MonsterLLM(model="microsoft/Phi-3-mini-4k-instruct")
    
        response = llm.complete("What is the capital of France?")
        print(str(response))
        ```
    

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llama_index_llms_monsterapi-0.4.1.tar.gz (5.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

llama_index_llms_monsterapi-0.4.1-py3-none-any.whl (5.2 kB view details)

Uploaded Python 3

File details

Details for the file llama_index_llms_monsterapi-0.4.1.tar.gz.

File metadata

File hashes

Hashes for llama_index_llms_monsterapi-0.4.1.tar.gz
Algorithm Hash digest
SHA256 c9a04543173ee89cee5bc1588d1ed2ba7d98dae8ee947d0a47e7ba7776abe9c1
MD5 f4de8e48c27de767c9aee96f0e7efa56
BLAKE2b-256 76a96d9fd0eec2bfbffd42894714e9b8df0166a60c7b804bd54842eeaa4498cb

See more details on using hashes here.

File details

Details for the file llama_index_llms_monsterapi-0.4.1-py3-none-any.whl.

File metadata

File hashes

Hashes for llama_index_llms_monsterapi-0.4.1-py3-none-any.whl
Algorithm Hash digest
SHA256 851f20920f10cc54b54b285c188f590ba3f0384a876694829370ac6d061cc1b2
MD5 7193091dc7bbd27c608ebc7625426bf7
BLAKE2b-256 26f506a36a1f1a8d7b9f38fa9ec3083b5066316641b3bda8e961f382a7f0c7ef

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page