Skip to main content

aify is an AI-native application framework and runtime

Project description

🚀 aify

Build your AI-native application in seconds.

Home | Documentation | Feedback

🛠️ AI-native application framework and runtime. Simply write a YAML file.

🤖 Ready-to-use AI chatbot UI.

Dependencies

Features

  • Models: The LLMs/transformers models supported by guidance.
  • Memory storage: Local file / Google Cloud Datastore / User-defined
  • Embeddings: OpenAI / User-defined
  • Vector storage and search: Local CSV files, Pandas DataFrame and Numpy in memory / User-defined
  • Deployment: Local / Google Cloud App engine
  • UI: Chatbot webui
  • API: RESTful API / Python

Getting started

Welcome to Aify, the AI-native application framework and runtime that allows you to ship your AI applications in seconds! With Aify, you can easily build and deploy AI-powered applications using a simple YAML file. In this guide, we will walk you through the steps to get started with Aify and create your first AI application.

Installation

To begin, make sure you have the following prerequisites installed on your system:

  • Python 3.8 or higher
  • Pip package manager

Once you have the prerequisites, you can install Aify by running the following command in your terminal:

pip install aify

Create your first app

You need to prepare a directory for your applications:

mkdir ./apps

Now you can start the aify service and then access http://localhost:2000 using a browser, and aify will greet you.

aify run ./apps

aify screenshot

Now it's just a blank application, you can't use it for anything. Next, we will create a chatbot.

Creating a YAML file aify uses a YAML file to define your AI application. This file contains all the necessary configurations and settings for your application. Here's an example of a basic YAML file:

title: Chatbot

model:
  vendor: openai
  name: gpt-3.5-turbo
  params:
    api_key: <YOUR_OPENAI_API_KEY>

prompt: |
  {{#system~}}
  You are a helpful and terse assistant.
  {{~/system}}

  {{#each (memory.read program_name session_id n=3)}}
  {{~#if this.role == 'user'}}
  {{#user~}}
  {{this.content}}
  {{~/user}}
  {{/if~}}
  {{~#if this.role == 'assistant'}}
  {{#assistant~}}
  {{this.content}}
  {{~/assistant}}
  {{/if~}}
  {{~/each}}

  {{#user~}}
  {{prompt}}
  {{memory.save program_name session_id 'user' prompt}}
  {{~/user}}

  {{#assistant~}}
  {{gen 'answer' temperature=0 max_tokens=2000}}
  {{memory.save program_name session_id 'assistant' answer}}
  {{~/assistant}}

variables:
  - name: prompt
    type: input
  - name: answer
    type: output

Here are some simple explanations about this YAML file:

  • The title represents the name of this application.
  • The model section defines the AI model used by this application and the runtime parameters required by the model.
  • The prompt section is used to drive the application's execution. Aify uses the guidance software package provided by Microsoft to drive the execution of the AI program. Guidance provides a way to operate as a Chain of Thought. Since guidance uses the Handlebars template system, the format of this section is actually a Handlebars template.The prompt section contains some helper functions that allow the AI model to dynamically change its runtime behavior, helping us achieve more complex functionality. These functions are built-in to aify, but you can also write your own helper functions in Python to accomplish specific tasks.
    • The terms "system," "user," and "assistant" are used to define the roles in an LLM-based chat task.
    • "memory.read" and "memory.write" are built-in helper functions in Aify, used to save and load the conversation history of users and AI.
    • "each" and "if" are branch control statements provided by Handlebars.
    • "gen" is the function provided by "guidance" to indicate the execution of LLM generation tasks.
  • The variables section defines the input and output variables of the application, which are used for external systems to access the data generated by AI through an API.

Play with your AI app

Now go back to your browser and refresh the page. You will see the application you just created. You can have some conversations with it, just like ChatGPT.

aify screenshot

aify is not a chatbot

Although aify provides a chatbot interface, its main purpose is not to provide a replacement for ChatGPT or a competitive conversation application.

The chatbot UI is only for convenient debugging of AI applications. Of course, you can indeed use it as a chatbot for daily use.

The main goal of aify is to provide an efficient framework for developing and deploying AI applications.

If your goal is to develop your own complex AI applications, you should pay more attention to the APIs and extension mechanisms provided by aify.

📝 More examples: https://github.com/shellc/aify/tree/main/examples

Webui screenshot

Webui screenshot

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distribution

aify-0.1.23-py3-none-any.whl (1.4 MB view details)

Uploaded Python 3

File details

Details for the file aify-0.1.23-py3-none-any.whl.

File metadata

  • Download URL: aify-0.1.23-py3-none-any.whl
  • Upload date:
  • Size: 1.4 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.3

File hashes

Hashes for aify-0.1.23-py3-none-any.whl
Algorithm Hash digest
SHA256 d3dd3760ac5ba7ecd1ff92787eb97afd00cb9365ff91e862f5de573cfaee657f
MD5 940bb2499ca898ed12a9f9b439089408
BLAKE2b-256 0596ce62d581c407dfe8781f68768ee2d00f5175318a22ae92291f80a56b087f

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page