Skip to main content

A package for building chatbots easily without needing of any high system resources.

Project description

WebraftChatBot

The "chatbot-webraft" Python module is a powerful tool that allows users to create their own chatbots with custom datasets in an easy and efficient manner. It is designed to work with CSV datasets and provides a variety of model types to choose from, ensuring that users can select the best model for their specific needs. The module requires no high-end servers and can be used with or without training, making it a flexible and cost-effective solution for chatbot development. With the "chatbot-webraft" module, users can quickly and easily develop their own chatbots, customizing the dataset and model type to fit their specific requirements. Whether you're a beginner or an experienced developer, the "chatbot-webraft" module provides a simple and intuitive way to create custom chatbots with Python due to its easy to use syntax.

One of the best features of Chatbot-Webraft is its ability to generate code scripts in specific programming languages from text. This means that you can use your preferred programming language, or even write your own code, and Chatbot-Webraft will convert it into a script that is ready to use. Whether you are an experienced developer or just starting out, this feature makes it easy to build chatbots quickly and efficiently.

In summary, Chatbot-Webraft is a must-have tool for anyone looking to create chatbots. With its wide range of features, including multiple model types, easy customization, and code generation capabilities, Chatbot-Webraft is the perfect tool for anyone looking to build chatbots with ease and efficiency.

PLEASE NOTE THAT THIS MODULE IS IN BETA AND CHATBOTS CREATED WITH IT MAY RESPOND WITH UNEXPECTED ANSWERS

USAGE:

Import library

from chatbot_webraft import chatbot

Create Model

chatbot.create_model("my-model")

Load CSV DATASET (here 'input' needs to be replaced with the column in dataset that has the fields containing the data to be inputted same to be done with 'label' but with fields containing the data to be outputted. )

chatbot.dataset(CSV_FILE_NAME,'input','label',"my-model"")

Add more data to chatbot(OPTIONAL)

chatbot.add_data("my-model"",["input1","input2"],["output1","output2"])

Specify Input

input = "hi" / input = input("Enter: ")

Run chatbot (for default set MODEL_TYPE to "spimx")

print(chatbot.model_load(MDDEL_TYPE,input,"my-model",MDDEL_TYPE))/chatbot.model_load(MDDEL_TYPE,input,"my-model",MDDEL_TYPE)

Use codewriter (currently python supported)

print(codewriter("python",input,"my-model","pywriter") / chatbot.model_load("pywriter",input,"my-model","pywriter")

Model Types:

Inbuilt model types: nlpm (Generative, Best) , spimx (better), spimxr (more good), spim (good), rasv (not that good)

You can use other model types by downloading them from here and define path where modeltype (.wrmodel) file is stored to MODEL_TYPE

NLPM: A generative model type (framework) , works best for generative bots . (BETA) NLPM USAGE: chatbot.model_load("nlpm",input,"my-model",MODELTYPE) - Replace MODELTYPE with the other modeltype that suites you the best , i.e - spim , spimx , rasv , spim (do not replace MODELTYPE with 'nlpm' rather replace with others)

There is also a transformer modified framework that uses Tokenizers from prebuilt transfomer model i.e - BERT . Note that it needs very high system for running and takes a lot of time to load. USE - chatbot.model_load("bert",input,"my-model","bert) (It doesnt need training)

Basic BOT Usage:

#Import library
from chatbot_webraft import chatbot

#set model name
model = "my-model" 

#create model
chatbot.create_model(model)

#load CSV dataset , Mention input column (question) and label column (answer)
chatbot.dataset(CSV_FILE_NAME,INPUT_COLUMN,LABEL_COLUMN,model) 


#run in loop
while True:
 prompt = input("You: ")    
 #run model and parse input
 print("Bot: ",chatbot.model_load("spimx",prompt,model,"spimx")) 

Bot usage for discord:

#Import libraries
from chatbot_webraft import chatbot
import discord 
client = discord.Client(intents=discord.Intents.all())


model = "my-model" #set model name
chatbot.create_model(model) #create model
chatbot.dataset(CSV_FILE_NAME,INPUT_COLUMN,LABEL_COLUMN,model) #load CSV dataset .. Mention input column (question) and label column (answer)
chatbot.add_data(model,["hey","what about you"],["hello bro","im fine ,you?"]) #add more data to model (Not saved , only in memory). Data that you may want to add later.
@client.event
async def on_message(message):
    if message.author == client.user:
        return

    prompt = message.content
    
    await message.reply(chatbot.model_load("spim",prompt,model,"spim")) #run model and parse output
client.run(BOT_TOKEN)

Basic codewriter usage:

 #Import library
from chatbot_webraft import chatbot

#set model name
model = "my-model" 

#create model
chatbot.create_model(model)

#load CSV dataset , Mention input column (question) and label column (answer)
chatbot.dataset(CSV_FILE_NAME,INPUT_COLUMN,LABEL_COLUMN,model) 


#run in loop
while True:
  prompt = input("You: ")    
  #run model and parse input
  print("Bot: ",chatbot.model_load("pywriter",prompt,model,"pywriter")) #Or codewriter("python",prompt,model,"pywriter")

Train Your own Model

In versions < 1.0.0 you can now train / finetune your own language models through minimal code.

Here is a basic example of pre training on chat data (using language data is advised): Before trying, download the sample dataset given below. (after the code)

from chatbot_webraft.yourmodel import TransformerModel
import torch

d_model = 128 # Adjust according to your needs and system specs

dataset_name = "sample.csv"  # Replace with your correct dataset path

input_set = "input"

label_set = "label"

batch_size = 16 # Adjust according to your needs and system specs

heads = 16 # Adjust according to your needs and system specs

num_layers = 6 # Adjust according to your needs and system specs

device = torch.device("cuda" if torch.cuda.is_available() else "cpu")

epochs = 40  

path = "path/to/folder/" # Eg. D:/Softwares/AI/Trainer/

model_name= path+"yourmodelname" # Eg. chatllm
transformer_model = TransformerModel(d_model, dataset_name, input_set, label_set, batch_size, heads, num_layers, device, epochs, model_name, max_len=256)
transformer_model.train() # Finally training

question = "hi"

print(transformer_model.inference(question)) # Doing Inference

You can also do inference already trained model but you require to initialise transformer model with that checkpoint and use same parameters (i.e. d_model, heads, num_layers, max_len)

checkpoint = torch.load(path+model_name+"_final.pth",map_location=device)
transformer = checkpoint['transformer']
transformer_model = TransformerModel(d_model, dataset_name, input_set, label_set, batch_size, heads, num_layers, device, epochs, model_name, max_len=256,transformer=transfromer)

You can also fine tune the model by just changing the dataset_name at initialising of new model and doing transformer_model.finetune().

SAMPLE CSV DATASET:

Download Sample csv dataset from here , Dont forget to put the file in your project directory
Change CSV_FILE_NAME to "sample.csv" , INPUT_COLUMN to "input" , LABEL_COLUMN to "label" in the code

Errors:

Error 1 - This error indicates that the model name used in specific function is not same as the modelname set with chatbot.createmodel()
Error 2 - This error indicates that the codewriter couldn't convert specific text to code due to lack of resources
Error 3 - This error indicates that the file being set in any of these functions: chatbot.dataset() , chatbot.model_load() doesn't exist

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

chatbot_webraft-1.0.2.tar.gz (124.2 kB view hashes)

Uploaded Source

Built Distribution

chatbot_webraft-1.0.2-py3-none-any.whl (127.9 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page