Skip to main content

An integration package connecting Prolog and LangChain

Project description

License: MIT PyPI version Python 3.10+ Documentation Status Version

LangChain-Prolog

A Python library that integrates SWI-Prolog with LangChain. It enables seamless blending of Prolog’s logic programming capabilities into LangChain applications, allowing rule-based reasoning, knowledge representation, and logical inference alongside GenAI models.

Features

  • Seamless integration between LangChain and SWI-Prolog
  • Use Prolog queries as LangChain's runnables and tools
  • Invoke Prolog predicates from LangChain's LLM models, chains and agents
  • Support for both synchronous and asynchronous operations
  • Comprehensive error handling and logging
  • Cross-platform support (macOS, Linux, Windows)

Installation

Prerequisites

  • Python 3.10 or later
  • SWI-Prolog installed on your system
  • The following Python libraries will be installed:
    • langchain 0.3.0 or later
    • janus-swi 1.5.0 or later
    • pydantic 2.0.0 or later

Once SWI-Prolog has been installed, langchain-prolog can be installed using pip:

pip install langchain-prolog

Runnable Interface

The PrologRunnable class allows the generation of langchain runnables that use Prolog rules to generate answers.

Let's use the following set of Prolog rules in the file family.pl:

parent(john, bianca, mary).
parent(john, bianca, michael).
parent(peter, patricia, jennifer).
partner(X, Y) :- parent(X, Y, _).

There are three ways to use a PrologRunnable to query Prolog:

1) Using a Prolog runnable with a full predicate string

from langchain_prolog import PrologConfig, PrologRunnable

config = PrologConfig(rules_file="family.pl")
prolog = PrologRunnable(prolog_config=config)
result = prolog.invoke("partner(X, Y)")
print(result)

We can pass a string representing a single predicate query. The invoke method will return True, False or a list of dictionaries with all the solutions to the query:

[{'X': 'john', 'Y': 'bianca'},
 {'X': 'john', 'Y': 'bianca'},
 {'X': 'peter', 'Y': 'patricia'}]

2) Using a Prolog runnable with a default predicate

from langchain_prolog import PrologConfig, PrologRunnable

config = PrologConfig(rules_file="family.pl", default_predicate="partner")
prolog = PrologRunnable(prolog_config=config)
result = prolog.invoke("peter, X")
print(result)

When using a default predicate, only the arguments for the predicate are passed to the Prolog runable, as a single string. Following Prolog conventions, uppercase identifiers are variables and lowercase identifiers are values (atoms or strings):

[{'X': 'patricia'}]

3) Using a Prolog runnable with a dictionary and schema validation

from langchain_prolog import PrologConfig, PrologRunnable

schema = PrologRunnable.create_schema("partner", ["man", "woman"])
config = PrologConfig(rules_file="family.pl", query_schema=schema)
prolog = PrologRunnable(prolog_config=config)
result = prolog.invoke({"man": None, "woman": "bianca"})
print(result)

If a schema is defined, we can pass a dictionary using the names of the parameters in the schema as the keys in the dictionary. The values can represent Prolog variables (uppercase first letter) or strings (lower case first letter). A None value is interpreted as a variable and replaced with the key capitalized:

[{'Man': 'john'}, {'Man': 'john'}]

You can also pass a Pydantic object generated with the schema to the invoke method:

args = schema(man='M', woman='W')
result = prolog.invoke(args)
print(result)

Uppercase values are treated as variables:

[{'M': 'john', 'W': 'bianca'},
 {'M': 'john', 'W': 'bianca'},
 {'M': 'peter', 'W': 'patricia'}]

Tool Interface

The PrologTool class allows the generation of langchain tools that use Prolog rules to generate answers.

See the Runnable Interface section for the conventions on hown to pass variables and values to the Prolog interpreter.

Let's use the following set of Prolog rules in the file family.pl:

parent(john, bianca, mary).
parent(john, bianca, michael).
parent(peter, patricia, jennifer).
partner(X, Y) :- parent(X, Y, _).

There are three diferent ways to use a PrologTool to query Prolog:

1) Using a Prolog tool with an LLM and function calling

First create the Prolog tool:

from langchain_prolog import PrologConfig, PrologRunnable, PrologTool

schema = PrologRunnable.create_schema("parent", ["man", "woman", "child"])
config = PrologConfig(
    rules_file="family.pl",
    query_schema=schema,
)
prolog_tool = PrologTool(
    prolog_config=config,
    name="family_query",
    description="""
        Query family relationships using Prolog.
        parent(X, Y, Z) implies only that Z is a child of X and Y.
        Input must be a dictionary like:
        {
            'man': 'richard',
            'woman': 'valery',
            'child': None,
        
        }
        Use 'None' to indicate a variable that need to be instantiated by Prolog
        The query will return:
            - 'True': if the relationship 'child' is a child of 'men' and 'women' holds.
            - 'False' if the relationship 'child' is a child of 'man' and 'woman' does not holds.
            - A list of dictionaries with all the answers to the Prolog query
        Do not use double quotes.
    """,
)

Then bind it to the LLM model and query the model:

from langchain_openai import ChatOpenAI
from langchain_core.messages import HumanMessage

llm = ChatOpenAI(model="gpt-4o-mini")
llm_with_tools = llm.bind_tools([prolog_tool])
messages = [HumanMessage("Who are John's children?")]
response = llm_with_tools.invoke(messages)
messages.append(response)
print(response.tool_calls[0])

The LLM will respond with a tool call request:

{'name': 'family_query',
 'args': {'man': 'john', 'woman': None, 'child': None},
 'id': 'call_V6NUsJwhF41G9G7q6TBmghR0',
 'type': 'tool_call'}

The tool takes this request and queries the Prolog database:

tool_msg = prolog_tool.invoke(response.tool_calls[0])
messages.append(tool_msg)
print(tool_msg)

The tool returns a list with all the solutions for the query:

content='[{"Woman": "bianca", "Child": "mary"}, {"Woman": "bianca", "Child": "michael"}]'
name='family_query'
tool_call_id='call_V6NUsJwhF41G9G7q6TBmghR0'

That we then pass to the LLM:

answer = llm_with_tools.invoke(messages)
print(answer.content)

And the LLM answers the original query using the tool response:

John has two children: Mary and Michael. Their mother is Bianca.

2) Using a Prolog tool with a LangChain agent

First create the Prolog tool:

from langchain_prolog import PrologConfig, PrologRunnable, PrologTool

schema = PrologRunnable.create_schema("parent", ["man", "woman", "child"])
config = PrologConfig(
    rules_file="family.pl",
    query_schema=schema,
)
prolog_tool = PrologTool(
    prolog_config=config,
    name="family_query",
    description="""
        Query family relationships using Prolog.
        parent(X, Y, Z) implies only that Z is a child of X and Y.
        Input must be a dictionary like:
        {
            'man': 'richard',
            'woman': 'valery',
            'child': None,
        
        }
        Use 'None' to indicate a variable that need to be instantiated by Prolog
        The query will return:
            - 'True': if the relationship 'child' is a child of 'men' and 'women' holds.
            - 'False' if the relationship 'child' is a child of 'man' and 'woman' does not holds.
            - A list of dictionaries with all the answers to the Prolog query
        Do not use double quotes.
    """,
)

Then pass it to the agent's constructor:

from langchain_core.prompts import ChatPromptTemplate
from langchain.agents import create_tool_calling_agent, AgentExecutor

llm = ChatOpenAI(model="gpt-4o-mini")
prompt = ChatPromptTemplate.from_messages(
    [
        ("system", "You are a helpful assistant"),
        ("human", "{input}"),
        ("placeholder", "{agent_scratchpad}"),
    ]
)
tools = [prolog_tool]
agent = create_tool_calling_agent(llm, tools, prompt)
agent_executor = AgentExecutor(agent=agent, tools=tools)

The agent takes the query and use the Prolog tool if needed:

answer = agent_executor.invoke({"input": "Who are John's children?"})
print(answer["output"])

Then the agent recieves the tool response as part of the {agent_scratchpad} placeholder and generates the answer:

John has two children: Mary and Michael. Their mother is Bianca.

3) Using a Prolog tool with a LangGraph agent

First create the Prolog tool:

from langchain_prolog import PrologConfig, PrologRunnable, PrologTool

schema = PrologRunnable.create_schema("parent", ["man", "woman", "child"])
config = PrologConfig(
    rules_file="family.pl",
    query_schema=schema,
)
prolog_tool = PrologTool(
    prolog_config=config,
    name="family_query",
    description="""
        Query family relationships using Prolog.
        parent(X, Y, Z) implies only that Z is a child of X and Y.
        Input must be a dictionary like:
        {
            'man': 'richard',
            'woman': 'valery',
            'child': None,
        
        }
        Use 'None' to indicate a variable that need to be instantiated by Prolog
        The query will return:
            - 'True': if the relationship 'child' is a child of 'men' and 'women' holds.
            - 'False' if the relationship 'child' is a child of 'man' and 'woman' does not holds.
            - A list of dictionaries with all the answers to the Prolog query
        Do not use double quotes.
    """,
)

Then pass it to the agent's constructor:

from langgraph.prebuilt import create_react_agent

lg_agent = create_react_agent(llm, [prolog_tool])

The agent takes the query and use the Prolog tool if needed:

messages =lg_agent.invoke({"messages": [("human", query)]})
messages["messages"][-1].pretty_print()

Then the agent receives​ the tool response and generates the answer:

================================== Ai Message ==================================

John has two children: Mary and Michael, with Bianca as their mother.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

langchain_prolog-0.1.1.post17.tar.gz (19.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

langchain_prolog-0.1.1.post17-py3-none-any.whl (17.9 kB view details)

Uploaded Python 3

File details

Details for the file langchain_prolog-0.1.1.post17.tar.gz.

File metadata

  • Download URL: langchain_prolog-0.1.1.post17.tar.gz
  • Upload date:
  • Size: 19.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/2.3.2 CPython/3.10.19 Linux/6.14.0-1017-azure

File hashes

Hashes for langchain_prolog-0.1.1.post17.tar.gz
Algorithm Hash digest
SHA256 f434c9baa143bdf4a49d5c1c83a992b737659c54a78f01dc08ad13b3caf8f4d2
MD5 d14ea51bfebbd4cebd92e9ffa63a37aa
BLAKE2b-256 2751bb465031e981eceaa2e12978f8d561b7c861e5be50cc319f04a85643022b

See more details on using hashes here.

File details

Details for the file langchain_prolog-0.1.1.post17-py3-none-any.whl.

File metadata

File hashes

Hashes for langchain_prolog-0.1.1.post17-py3-none-any.whl
Algorithm Hash digest
SHA256 674f45c42dff1457cb07250c217a9d64d3e51d485edc1f16ba8c5810447f762f
MD5 39d25ed67498de8f1267444519d6b3f7
BLAKE2b-256 a654f1962fb39fc822fead2ac7cb836924c4ed1e63f1d45b5682b0912bc808e8

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page