Skip to main content

simple flask server to host OpenVoiceOS persona plugins as a service

Project description

Persona Server

Running

$ ovos-persona-server --persona rivescript_bot.json

Personas

personas don't need to use LLMs, you don't need a beefy GPU to use ovos-persona, find solver plugins here

some repos and skills also provide solvers, such as ovos-classifiers (wordnet), skill-ddg, skill-wikipedia and skill-wolfie

{
  "name": "OldSchoolBot",
  "solvers": [
    "ovos-solver-wikipedia-plugin",
    "ovos-solver-ddg-plugin",
    "ovos-solver-plugin-wolfram-alpha",
    "ovos-solver-wordnet-plugin",
    "ovos-solver-rivescript-plugin",
    "ovos-solver-failure-plugin"
  ],
  "ovos-solver-plugin-wolfram-alpha": {"appid": "Y7353-9HQAAL8KKA"}
}

this persona would search ddg api / wikipedia for "what is"/"tell me about" questions, falling back to wordnet when offline for dictionary look up, and finally rivescript for general chitchat, we also add the failure solver to be sure the persona always says something

wolfram alpha illustrates how to pass solver configs, it has a requirement for an API key

search/knowledge base solvers can be used together with LLM solvers to ensure factual answers and act as a tool/internet access layer, in the example above you would typically replace rivescript with a LLM.

Some solvers may also use other solvers internally, such as a MOS (Mixture Of Solvers)

Client side usage

OpenAI compatible API, for usage with OVOS see ovos-solver-plugin-openai-persona

import openai

openai.api_key = ""
openai.api_base = "http://localhost:8337"

# NOTE - most solvers don't support a chat history,
#  only last message in messages list is considered
chat_completion = openai.ChatCompletion.create(
    model="",  # individual personas might support this, passed under context
    messages=[{"role": "user", "content": "tell me a joke"}],
    stream=False,
)

if isinstance(chat_completion, dict):
    # not stream
    print(chat_completion.choices[0].message.content)
else:
    # stream
    for token in chat_completion:
        content = token["choices"][0]["delta"].get("content")
        if content != None:
            print(content, end="", flush=True)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

ovos-persona-server-0.3.1a1.tar.gz (8.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

ovos_persona_server-0.3.1a1-py3-none-any.whl (9.1 kB view details)

Uploaded Python 3

File details

Details for the file ovos-persona-server-0.3.1a1.tar.gz.

File metadata

  • Download URL: ovos-persona-server-0.3.1a1.tar.gz
  • Upload date:
  • Size: 8.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.9.21

File hashes

Hashes for ovos-persona-server-0.3.1a1.tar.gz
Algorithm Hash digest
SHA256 470bc45335284a09efd1ef11c5b1974649e9d551cb39dabb41dfecbc523e7fdf
MD5 e979c6d0e328c7b9fbc1dd6a7218c85c
BLAKE2b-256 b382af2582350a70e96de44fe1793bdd6286a84e3db3343fc3171f9d9d990633

See more details on using hashes here.

File details

Details for the file ovos_persona_server-0.3.1a1-py3-none-any.whl.

File metadata

File hashes

Hashes for ovos_persona_server-0.3.1a1-py3-none-any.whl
Algorithm Hash digest
SHA256 13ee94a06d83f1c32637aba7a036824826fcb37c5db1b600d5d8e03d901dc799
MD5 b477cc76035dd35c4e7944b5293b1ac4
BLAKE2b-256 e71c8badfa8368d4bc4daa4bce19dfae5f45079e1d5217e1d6c8600883da0ce5

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page