Skip to main content

An integration package connecting Prolog and LangChain

Project description

License: MIT PyPI version Python 3.10+ Documentation Status Version

LangChain-Prolog

A Python library that integrates SWI-Prolog with LangChain. It enables seamless blending of Prolog’s logic programming capabilities into LangChain applications, allowing rule-based reasoning, knowledge representation, and logical inference alongside GenAI models.

Features

  • Seamless integration between LangChain and SWI-Prolog
  • Use Prolog queries as LangChain's runnables and tools
  • Invoke Prolog predicates from LangChain's LLM models, chains and agents
  • Support for both synchronous and asynchronous operations
  • Comprehensive error handling and logging
  • Cross-platform support (macOS, Linux, Windows)

Installation

Prerequisites

  • Python 3.10 or later
  • SWI-Prolog installed on your system
  • The following Python libraries will be installed:
    • langchain 0.3.0 or later
    • janus-swi 1.5.0 or later
    • pydantic 2.0.0 or later

Once SWI-Prolog has been installed, langchain-prolog can be installed using pip:

pip install langchain-prolog

Runnable Interface

The PrologRunnable class allows the generation of langchain runnables that use Prolog rules to generate answers.

Let's use the following set of Prolog rules in the file family.pl:

parent(john, bianca, mary).
parent(john, bianca, michael).
parent(peter, patricia, jennifer).
partner(X, Y) :- parent(X, Y, _).

There are three ways to use a PrologRunnable to query Prolog:

1) Using a Prolog runnable with a full predicate string

from langchain_prolog import PrologConfig, PrologRunnable

config = PrologConfig(rules_file="family.pl")
prolog = PrologRunnable(prolog_config=config)
result = prolog.invoke("partner(X, Y)")
print(result)

We can pass a string representing a single predicate query. The invoke method will return True, False or a list of dictionaries with all the solutions to the query:

[{'X': 'john', 'Y': 'bianca'},
 {'X': 'john', 'Y': 'bianca'},
 {'X': 'peter', 'Y': 'patricia'}]

2) Using a Prolog runnable with a default predicate

from langchain_prolog import PrologConfig, PrologRunnable

config = PrologConfig(rules_file="family.pl", default_predicate="partner")
prolog = PrologRunnable(prolog_config=config)
result = prolog.invoke("peter, X")
print(result)

When using a default predicate, only the arguments for the predicate are passed to the Prolog runable, as a single string. Following Prolog conventions, uppercase identifiers are variables and lowercase identifiers are values (atoms or strings):

[{'X': 'patricia'}]

3) Using a Prolog runnable with a dictionary and schema validation

from langchain_prolog import PrologConfig, PrologRunnable

schema = PrologRunnable.create_schema("partner", ["man", "woman"])
config = PrologConfig(rules_file="family.pl", query_schema=schema)
prolog = PrologRunnable(prolog_config=config)
result = prolog.invoke({"man": None, "woman": "bianca"})
print(result)

If a schema is defined, we can pass a dictionary using the names of the parameters in the schema as the keys in the dictionary. The values can represent Prolog variables (uppercase first letter) or strings (lower case first letter). A None value is interpreted as a variable and replaced with the key capitalized:

[{'Man': 'john'}, {'Man': 'john'}]

You can also pass a Pydantic object generated with the schema to the invoke method:

args = schema(man='M', woman='W')
result = prolog.invoke(args)
print(result)

Uppercase values are treated as variables:

[{'M': 'john', 'W': 'bianca'},
 {'M': 'john', 'W': 'bianca'},
 {'M': 'peter', 'W': 'patricia'}]

Tool Interface

The PrologTool class allows the generation of langchain tools that use Prolog rules to generate answers.

See the Runnable Interface section for the conventions on hown to pass variables and values to the Prolog interpreter.

Let's use the following set of Prolog rules in the file family.pl:

parent(john, bianca, mary).
parent(john, bianca, michael).
parent(peter, patricia, jennifer).
partner(X, Y) :- parent(X, Y, _).

There are three diferent ways to use a PrologTool to query Prolog:

1) Using a Prolog tool with an LLM and function calling

First create the Prolog tool:

from langchain_prolog import PrologConfig, PrologRunnable, PrologTool

schema = PrologRunnable.create_schema("parent", ["man", "woman", "child"])
config = PrologConfig(
    rules_file="family.pl",
    query_schema=schema,
)
prolog_tool = PrologTool(
    prolog_config=config,
    name="family_query",
    description="""
        Query family relationships using Prolog.
        parent(X, Y, Z) implies only that Z is a child of X and Y.
        Input must be a dictionary like:
        {
            'man': 'richard',
            'woman': 'valery',
            'child': None,
        
        }
        Use 'None' to indicate a variable that need to be instantiated by Prolog
        The query will return:
            - 'True': if the relationship 'child' is a child of 'men' and 'women' holds.
            - 'False' if the relationship 'child' is a child of 'man' and 'woman' does not holds.
            - A list of dictionaries with all the answers to the Prolog query
        Do not use double quotes.
    """,
)

Then bind it to the LLM model and query the model:

from langchain_openai import ChatOpenAI
from langchain_core.messages import HumanMessage

llm = ChatOpenAI(model="gpt-4o-mini")
llm_with_tools = llm.bind_tools([prolog_tool])
messages = [HumanMessage("Who are John's children?")]
response = llm_with_tools.invoke(messages)
messages.append(response)
print(response.tool_calls[0])

The LLM will respond with a tool call request:

{'name': 'family_query',
 'args': {'man': 'john', 'woman': None, 'child': None},
 'id': 'call_V6NUsJwhF41G9G7q6TBmghR0',
 'type': 'tool_call'}

The tool takes this request and queries the Prolog database:

tool_msg = prolog_tool.invoke(response.tool_calls[0])
messages.append(tool_msg)
print(tool_msg)

The tool returns a list with all the solutions for the query:

content='[{"Woman": "bianca", "Child": "mary"}, {"Woman": "bianca", "Child": "michael"}]'
name='family_query'
tool_call_id='call_V6NUsJwhF41G9G7q6TBmghR0'

That we then pass to the LLM:

answer = llm_with_tools.invoke(messages)
print(answer.content)

And the LLM answers the original query using the tool response:

John has two children: Mary and Michael. Their mother is Bianca.

2) Using a Prolog tool with a LangChain agent

First create the Prolog tool:

from langchain_prolog import PrologConfig, PrologRunnable, PrologTool

schema = PrologRunnable.create_schema("parent", ["man", "woman", "child"])
config = PrologConfig(
    rules_file="family.pl",
    query_schema=schema,
)
prolog_tool = PrologTool(
    prolog_config=config,
    name="family_query",
    description="""
        Query family relationships using Prolog.
        parent(X, Y, Z) implies only that Z is a child of X and Y.
        Input must be a dictionary like:
        {
            'man': 'richard',
            'woman': 'valery',
            'child': None,
        
        }
        Use 'None' to indicate a variable that need to be instantiated by Prolog
        The query will return:
            - 'True': if the relationship 'child' is a child of 'men' and 'women' holds.
            - 'False' if the relationship 'child' is a child of 'man' and 'woman' does not holds.
            - A list of dictionaries with all the answers to the Prolog query
        Do not use double quotes.
    """,
)

Then pass it to the agent's constructor:

from langchain_core.prompts import ChatPromptTemplate
from langchain.agents import create_tool_calling_agent, AgentExecutor

llm = ChatOpenAI(model="gpt-4o-mini")
prompt = ChatPromptTemplate.from_messages(
    [
        ("system", "You are a helpful assistant"),
        ("human", "{input}"),
        ("placeholder", "{agent_scratchpad}"),
    ]
)
tools = [prolog_tool]
agent = create_tool_calling_agent(llm, tools, prompt)
agent_executor = AgentExecutor(agent=agent, tools=tools)

The agent takes the query and use the Prolog tool if needed:

answer = agent_executor.invoke({"input": "Who are John's children?"})
print(answer["output"])

Then the agent recieves the tool response as part of the {agent_scratchpad} placeholder and generates the answer:

John has two children: Mary and Michael. Their mother is Bianca.

3) Using a Prolog tool with a LangGraph agent

First create the Prolog tool:

from langchain_prolog import PrologConfig, PrologRunnable, PrologTool

schema = PrologRunnable.create_schema("parent", ["man", "woman", "child"])
config = PrologConfig(
    rules_file="family.pl",
    query_schema=schema,
)
prolog_tool = PrologTool(
    prolog_config=config,
    name="family_query",
    description="""
        Query family relationships using Prolog.
        parent(X, Y, Z) implies only that Z is a child of X and Y.
        Input must be a dictionary like:
        {
            'man': 'richard',
            'woman': 'valery',
            'child': None,
        
        }
        Use 'None' to indicate a variable that need to be instantiated by Prolog
        The query will return:
            - 'True': if the relationship 'child' is a child of 'men' and 'women' holds.
            - 'False' if the relationship 'child' is a child of 'man' and 'woman' does not holds.
            - A list of dictionaries with all the answers to the Prolog query
        Do not use double quotes.
    """,
)

Then pass it to the agent's constructor:

from langgraph.prebuilt import create_react_agent

lg_agent = create_react_agent(llm, [prolog_tool])

The agent takes the query and use the Prolog tool if needed:

messages =lg_agent.invoke({"messages": [("human", query)]})
messages["messages"][-1].pretty_print()

Then the agent receives​ the tool response and generates the answer:

================================== Ai Message ==================================

John has two children: Mary and Michael, with Bianca as their mother.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

langchain_prolog-0.1.1.post11.tar.gz (19.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

langchain_prolog-0.1.1.post11-py3-none-any.whl (17.8 kB view details)

Uploaded Python 3

File details

Details for the file langchain_prolog-0.1.1.post11.tar.gz.

File metadata

  • Download URL: langchain_prolog-0.1.1.post11.tar.gz
  • Upload date:
  • Size: 19.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/2.1.3 CPython/3.10.18 Linux/6.11.0-1015-azure

File hashes

Hashes for langchain_prolog-0.1.1.post11.tar.gz
Algorithm Hash digest
SHA256 d4824f59508aa398c4b208fdfe6b10c1721b99ea0e89796206cafc34f68c166c
MD5 1ea1c7d70b76e93e211f89ba2db7327e
BLAKE2b-256 9c3402128fa457ea83666937efce39e65eeb31976c117cbec3b423199bf12f09

See more details on using hashes here.

File details

Details for the file langchain_prolog-0.1.1.post11-py3-none-any.whl.

File metadata

File hashes

Hashes for langchain_prolog-0.1.1.post11-py3-none-any.whl
Algorithm Hash digest
SHA256 a1b30b59cf329b5aa18747bc524deb5bc5577b7a7d6511881bad8eefbeec3f26
MD5 151ec7684e43c234390bd64d0ca5a8a2
BLAKE2b-256 9c279e079ec1389de22f7d20d8b2338f120251598a82773a043b209757902380

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page