Skip to main content

Mindcraft: an LLM-based engine for creating real NPCs. Empowered by Hugging Face, quantized LLMs with AWQ (thanks @TheBloke) and vLLM. It follows a RAG approach with chunk or sentence splitting, and a vector store. Right now, ChromaDB is the supported Vector Store and chunk splitting using `tiktoken` or sentence splitting using `spacy` are available.

Project description

mindcraft

MindCraft

The open-source NLP library to craft the minds of your NPC characters for your video games.

Requires Python 3.10 or higher.

It includes the following features:

  • Text generation using LLMs
  • Motivations, personality, personal goals
  • Knowledge and awareness about the world (RAG)
  • Short and Long-term memory (RAG)
  • Conversational styles
  • Supervised finetuning by human feedback (SFT)
  • Integration with vLLM for fast inference and streaming locally, remotelly or in Docker.
  • Usage of quantized AWQ models
  • Integration with API and RPC (to come!)

Create a World from a book

from mindcraft.infra.engine.llm_types import LLMType
from mindcraft.infra.vectorstore.stores_types import StoresTypes
from mindcraft.infra.embeddings.embeddings_types import EmbeddingsTypes
from mindcraft.lore.world import World

world = World(world_name="Middle Earth from the Lord of the Rings",
              embeddings=EmbeddingsTypes.MINILM,
              store_type=StoresTypes.CHROMA,
              llm_type=LLMType.YI_6B_AWQ,
              fast=False,      # <--- fast=True to switch on vLLM
              remote=False,    # <--- remote=True to use a remote vLLM server
              streaming=True) # <--- streaming=True to use vLLM streaming

Now we use some book to carry out chunk splitting and add it to our favourite Vector Database (in our case, ChromaDB)

from mindcraft.infra.splitters.text_splitters_types import TextSplitterTypes

world.book_to_world(book_path="/content/lotr1.txt",
                    text_splitter=TextSplitterTypes.SENTENCE_SPLITTER,
                    max_units=3,
                    overlap=1,
                    encoding='utf-8')

Query the lore of the world

Once a world has been created and populated with lore, query the lore known by NPCs by doing:

results = world.get_lore("What do you think about the Rings of Power?", num_results=5, min_similarity=0.95)
for i, d in enumerate(results.documents):
    print(f"SENTENCE {i}:\n{d}")
    print()

Instantiate an NPC in a world

Once a world has been created and populated with lore, instantiate an NPC in it by doing:

from mindcraft.features.motivation import Motivation
from mindcraft.features.personality import Personality
from mindcraft.features.mood import Mood

from mindcraft.mind.npc import NPC

name = "Galadriel"
description = "The Elven Queen of Lothlorien, bearer of Nenya, wife to Celeborn"
personalities = [Personality(x) for x in ['fair', 'mighty', 'wise', 'carying', 'kind']]
motivations = [Motivation(x) for x in ['Destroying the Evil', 'Protecting Middle Earth', 'Travelling West']]
mood = Mood('worried')

galadriel = NPC(name,
                description,
                personalities,
                motivations,
                mood,
                StoresTypes.CHROMA,
                EmbeddingsTypes.MINILM)

galadriel.add_npc_to_world()

Ask questions to the NPC

We get an iterator for the responses, to allow inference in streaming way.

answer_iter = galadriel.react_to("What do you think about the Rings of Power",
                               min_similarity=0.85,
                               ltm_num_results=3,
                               world_num_results=7,
                               max_tokens=600)

So your answers will be in the iterator, don't forget to loop through it!

for answer in answer_iter:
    print(answer)

Example of answer using Zephyr 7B quantized to 4b

Alas, my dear friend, the Rings of Power are a perilous burden that should not be wielded by those of mortal race. They possess a sinister force that overcomes the spirit, enslaving their very soul. I have observed their destructive potential in Eregion long ago, where many Elven-rings were created, each infused with its own unique potency. Some more potent than others, and often, they seem to have a propensity to gravitate towards the unsuspecting. In light of our shared concern for the wellbeing of Middle Earth, I implore you to heed my words; let us not succumb to the allure of these fateful rings, lest they consume us entirely.

Creating custom Supervised Finetuning datasets for NPCs

There are two loops integrated in the framework which allow you to create your own datasets.

  1. NPC.extract_conversational_styles_from_world: Retrieving conversations from the world and tagging the question and mood yourself
  2. NPC.extract_conversational_styles_talking_to_user: Creates conversations by talking to you

Supervised Finetuning

You can create your own datasets with supervised finetuning, accepting or rejecting the interactions. As a result, you will come up with a csv of interactions you can use to train or finetune your own models.

name||mood||question||answer
Galadriel||default||Good night, Galadriel!||'Good night, my friends! '\nsaid Galadriel. '\nSleep in peace!
Galadriel||grave||why he could say that?||....`He would be rash indeed that said that thing,' said Galadriel gravely.

LLM integrated

Quantized

  • TheBloke/mistral_7b_norobots-AWQ
  • TheBloke/zephyr-7B-beta-AWQ
  • TheBloke/notus-7B-v1-AWQ
  • TheBloke/Starling-LM-7B-alpha-AWQ
  • TheBloke/Yi-6B-AWQ
  • TheBloke/dragon-yi-6B-v0-AWQ

Unquantized

  • microsoft/phi-2
  • stabilityai/stablelm-zephyr-3b

Embeddings for RAG

CUDA and Torch in WINDOWS

If you are running on Windows on a machine with a GPU, and you get a message about not being able to find your gpu, you need to configure CUDA for Windows.

  1. Go to CUDA installation webpage.
  2. Select your Windows version and specifics.
  3. Download and install
  4. Uninstall torch (pip uninstall torch) 5a. Go to requirements.txt, comment the line after LINUX for torch and uncomment the line after WINDOWS 5b. Alternatively you can just run this command:
pip3 install torch -i https://download.pytorch.org/whl/cu121

You torch on windows CUDA should be working. To test it:

import torch

if __name__ == "__main__":
    print(torch.cuda.is_available())

vLLM

vLLM has been included for Fast Inference, in local, remote installations and Docker.

Local Fast inference (Paged Attention)

To use fast-inference, just run add fast=True to your World object:

world = World(world_name="Lord of the Rings",
              embeddings=EmbeddingsTypes.MINILM,
              store_type=StoresTypes.CHROMA,
              llm_type=LLMType.YI_6B_AWQ,
              fast=True) # <---- HERE

Remote Fast Inference

To the previous fast parameter, add also remote=True

world = World(world_name="Lord of the Rings",
              embeddings=EmbeddingsTypes.MINILM,
              store_type=StoresTypes.CHROMA,
              llm_type=LLMType.YI_6B_AWQ,
              fast=True,   # <---- HERE
              remote=True) # <---- HERE

Streaming (only if remote!)

To the previous fast and remote parameters, add also streaming=True

world = World(world_name="Lord of the Rings",
              embeddings=EmbeddingsTypes.MINILM,
              store_type=StoresTypes.CHROMA,
              llm_type=LLMType.YI_6B_AWQ,
              fast=True,   # <---- HERE
              remote=True,
              streaming=True) # <---- HERE

Example data

Notebooks

You can find notebooks in the notebooks folder of this project.

Demo 1: Creating a World and an NPC

Video

Architecture

mindcraft architecture

Tests

python -m pytest tests/*

Header

galadriel

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

mindcraft-0.2.4.tar.gz (31.7 kB view details)

Uploaded Source

Built Distribution

mindcraft-0.2.4-py3-none-any.whl (38.1 kB view details)

Uploaded Python 3

File details

Details for the file mindcraft-0.2.4.tar.gz.

File metadata

  • Download URL: mindcraft-0.2.4.tar.gz
  • Upload date:
  • Size: 31.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.10.4

File hashes

Hashes for mindcraft-0.2.4.tar.gz
Algorithm Hash digest
SHA256 6e69b7cdfabe82390a6ec12e33d900a896c351306394ada8748ddf8c17133a7a
MD5 a5a6a8a5270e6fe0a95bf9ba7eddd0b1
BLAKE2b-256 49709a970687303ad7e1c042f9dde5b3a24be591d088403407a453489a3b6c7d

See more details on using hashes here.

File details

Details for the file mindcraft-0.2.4-py3-none-any.whl.

File metadata

  • Download URL: mindcraft-0.2.4-py3-none-any.whl
  • Upload date:
  • Size: 38.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.10.4

File hashes

Hashes for mindcraft-0.2.4-py3-none-any.whl
Algorithm Hash digest
SHA256 620f83d0db5ee61ff03bfcc211d7b6cb6ca405d634a045a7817ab871b5abf8c9
MD5 2d3b541e728613ba09d1d06e7dd6a540
BLAKE2b-256 e1efb1ac006672c641a80086221792097e49ee18e9a6014f7504f641103d4330

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page