Skip to main content

chat to visualization with LLM

Project description

📈 Chat2Plot - interactive & safe text-to-visualization with LLM

This library uses LLMs to generate:

  1. Visualization
  2. High-level chart specifications in json (You can choose between a simple or vega-lite format)
  3. Explanation

from natural language requests for given data.

Chat2Plot does not generate executable code or SQL from the LLM, so you can safely generate visualizations.

demo: https://chat2plot-sample.streamlit.app/

sample

Quick Start

pip install chat2plot
import os
import pandas as pd
from chat2plot import chat2plot

# 1. Set api-key
os.environ["OPENAI_API_KEY"] = "..."

df = pd.read_csv(...)

# 2. Pass a dataframe to draw
c2p = chat2plot(df)

# 3. Make a question about the data
result = c2p("average target over countries")
result.figure.show()  # draw a plot
print(result.config)  # get a config (json / dataclass)
print(result.explanation)  # see the explanation generated by LLM

# you can make follow-up request to refine the chart
result = c2p("change to horizontal-bar chart")
result.figure.show()

Why Chat2Plot?

Inside Chat2Plot, LLM does not generate Python code, but generates plot specifications in json.

The declarative visualization specification in json is transformed into actual charts in Chat2Plot using plotly or altair, but users can also use json directly in their own applications.

This design limits the visualization expression compared to Python code generation (such as ChatGPT's Code Interpreter Plugin), but has the following practical advantages:

  • Secure

    • More secure execution, as LLM does not directly generate code.
  • Language-independent

    • Declarative data structures are language-agnostic, making it easy to plot in non-Python environments.
  • Interactive

    • Declarative data can be modified by the user to improve plots through collaborative work between the user and LLM.

Usage

A Chat2Plot instance can be created using the chat2plot function.

def chat2plot(
    df: pd.DataFrame,
    model_type: str = "simple",
    chat: langchain.chat_models.base.BaseChatModel | None = None,
    verbose: bool = False,
) -> Chat2PlotBase:
  • df - Data source for visualization.
  • model_type (optional) - Type of json format.
    • vega - A vega-lite compliant format
    • simple - A simpler format, parsed as chat2plot.PlotConfig
  • chat (optional) - The chat instance for interaction with LLMs. If omitted, ChatOpenAI(temperature=0, model_name="gpt-3.5-turbo") will be used.
  • verbose (optional) - If True, chat2plot will output logs.

Once an instance is created, a graph generation request can be made by calling query method, or simply passing the same arguments to the instance (__call__).

def query(self, q: str, config_only: bool = False, show_plot: bool = False) -> Plot:
  • q - A query string.
  • config_only - If True, skip generating figure.
  • show_plot - If True, showing the generated figure is also performed inside chat2plot.

The default behavior is config_only=False and show_plot=False, i.e. Chat2Plot generates a figure as well as configuration, but does not draw it.

The query method returns Plot object, and this has the following properties:

@dataclass(frozen=True)
class Plot:
    figure: alt.Chart | Figure | None
    config: PlotConfig | dict[str, Any] | None
    response_type: ResponseType
    explanation: str
    conversation_history: list[langchain.schema.BaseMessage] | None
  • figure - The generated figure (chart).
  • config - The chart specification. Returns dict for vega mode and PlotConfig for simple mode.
  • response_type - The result code of the response.
  • explanation - Reason for the chart setup, generated by LLMs.
  • conversation_history - Full history of conversation (including retry prompt).

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

chat2plot-0.0.10.tar.gz (14.7 kB view hashes)

Uploaded Source

Built Distribution

chat2plot-0.0.10-py3-none-any.whl (15.2 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page