Skip to main content

NEural-symbolic Entity Reasoning and Matching

Project description

Neer Match

neermatch website

License: MIT

The package neermatch provides a set of tools for neural-symbolic entity reasoning and matching. It is designed to support easy set-up, training, and inference of entity matching models using deep learning, symbolic learning, and a hybrid approach combining both deep and symbolic learning. Moreover, the package provides automated fuzzy logic reasoning (by refutation) functionality that can be used to examine the significance of particular associations between fields in an entity matching task.

The project is financially supported by the Deutsche Forschungsgemeinschaft (DFG) under Grant 539465691 as part of the Infrastructure Priority Programme “New Data Spaces for the Social Sciences” (SPP 2431).

The package has also an R implementation available at r-neer-match.

Features

The package is built on the concept of similarity maps. Similarity maps are concise representations of potential associations between fields in two datasets. Entities from two datasets can be matched using one or more pairs of fields (one from each dataset). Each field pair can have one or more ways to compute the similarity between the values of the fields.

Similarity maps are used to automate the construction of entity matching models and to facilitate the reasoning capabilities of the package. More details on the concept of similarity maps and an early implementation of the package’s functionality (without neural-symbolic components) are given by (Karapanagiotis and Liebald 2023).

The training loops for both deep and symbolic learning models are implemented in tensorflow (Abadi et al. 2015). The pure deep learning model inherits from the keras model class (Chollet et al. 2015). The neural-symbolic model is implemented using the logic tensor network (LTN) framework (Badreddine et al. 2022). Pure neural-symbolic and hybrid models do not inherit directly from the keras model class, but they emulate the behavior by providing custom compile, fit, evaluate, and predictmethods, so that all model classes in neermatch have a uniform calling interface.

Auxiliary Features

In addition, the package offers explainability functionality customized for the needs of matching problems. The default explainability behavior is built on the information provided by the similarity map. From a global explainability aspect, the package can be used to calculate partial matching dependencies and accumulated local effects on similarities. From a local explainability aspect, the package can be used to calculate local interpretable model-agnostic matching explanations and Shapley matching values.

Basic Usage

Implementing matching models using neermatch is a three-step process:

  1. Instantiate a model with a similarity map.
  2. Compile the model.
  3. Train the model.

To train the model you need to provide three datasets. Two datasets should contain records representing the entities to be matched. By convention, the first dataset is called Left and the second dataset is called Right dataset in the package’s documentation. The third dataset should contain the ground truth labels for the matching entities. The ground truth dataset should have two columns, one for the index of the entity in the Left dataset and one for the index of the entity in the Right dataset.

from neer_match.similarity_map import SimilarityMap
from neer_match.matching_model import NSMatchingModel
import tensorflow as tf

# 0) replace this with your own data preprocessing function
left, right, matches = prepare_data()

# 1) customize according to the fields in your data
smap = SimilarityMap(
    {
    "title": ["jaro_winkler"],
    "platform": ["levenshtein", "discrete"],
    "year": ["euclidean", "discrete"],
    "developer~dev": ["jaro"]
    }
)
model = NSMatchingModel(smap)

# 2) compile
model.compile(optimizer=tf.keras.optimizers.Adam(learning_rate=0.001))

# 3) train
model.fit(
    left, right, matches,
    epochs=51, batch_size=16,
    log_mod_n=10,
)
| Epoch      | BCE        | Recall     | Precision  | F1         | Sat        |
| 0          | 7.1111     | 1.0000     | 0.2500     | 0.4000     | 0.7330     |
| 10         | 6.4167     | 0.0000     | nan        | nan        | 0.7512     |
| 20         | 5.0065     | 0.0000     | nan        | nan        | 0.8265     |
| 30         | 3.6293     | 0.0000     | nan        | nan        | 0.8524     |
| 40         | 1.2699     | 1.0000     | 0.9750     | 0.9873     | 0.8937     |
| 50         | 0.6089     | 1.0000     | 0.9750     | 0.9873     | 0.9180     |
Training finished at Epoch 50 with DL loss 0.6089 and Sat 0.9180

Installation

From Source

You can obtain the sources for the development version of neermatch from its github repository.

git clone https://github.com/pi-kappa-devel/py-neer-match

To build and install the package locally, from the project’s root path, execute

python -m build
python -m pip install dist/$(basename `ls -Art dist | tail -n 1` -py3-none-any.whl).tar.gz

Documentation

Online documentation is available for the release version of the package.

Reproducing Documentation from Source

Make sure to build and install the package with the latest modifications before building the documentation. The documentation website is using sphinx. The build the documentation, from <project-root>/docs, execute

make html

Development Notes

Logo

The logo was designed using Microsoft Designer and GNU Image Manipulation Program (GIMP). The hexagon version of the logo was generated with the R package hexSticker. It uses the Philosopher font.

Alternative Software

TODO

Contributors

Pantelis Karapanagiotis (maintainer)

Marius Liebald (contributor)

Feel free to share, modify, and distribute. If you implement new features that might be of general interest, please consider contributing them back to the project.

License

The package is distributed under the MIT license.

References

Abadi, Martín, Ashish Agarwal, Paul Barham, Eugene Brevdo, Zhifeng Chen, Craig Citro, Greg S. Corrado, et al. 2015. “TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems.” https://www.tensorflow.org/.

Badreddine, Samy, Artur d’Avila Garcez, Luciano Serafini, and Michael Spranger. 2022. “Logic Tensor Networks.” Artificial Intelligence 303: 103649. https://doi.org/10.1016/j.artint.2021.103649.

Chollet, François et al. 2015. “Keras.” https://keras.io.

Karapanagiotis, Pantelis, and Marius Liebald. 2023. “Entity Matching with Similarity Encoding: A Supervised Learning Recommendation Framework for Linking (Big) Data.” http://dx.doi.org/10.2139/ssrn.4541376.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

neer_match-0.7.33.tar.gz (28.9 kB view details)

Uploaded Source

Built Distribution

neer_match-0.7.33-py3-none-any.whl (28.2 kB view details)

Uploaded Python 3

File details

Details for the file neer_match-0.7.33.tar.gz.

File metadata

  • Download URL: neer_match-0.7.33.tar.gz
  • Upload date:
  • Size: 28.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.10.15

File hashes

Hashes for neer_match-0.7.33.tar.gz
Algorithm Hash digest
SHA256 452e73dbafd160965ac2490cbb500781f5f2842deae97808a5fa3b76833c1722
MD5 6de399db959f848a9a130697661657e2
BLAKE2b-256 bc02bcc6bd0b499d38ef5092988e05bfeba67700e81427fe427aa3e99cefe6f3

See more details on using hashes here.

File details

Details for the file neer_match-0.7.33-py3-none-any.whl.

File metadata

  • Download URL: neer_match-0.7.33-py3-none-any.whl
  • Upload date:
  • Size: 28.2 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.10.15

File hashes

Hashes for neer_match-0.7.33-py3-none-any.whl
Algorithm Hash digest
SHA256 0ad80c7785fbcef2291ec5fb1340c4034cb34e9542142ad223a0f21c40344103
MD5 8241561a9e9c7e2bf6ce32ffe087f4a2
BLAKE2b-256 a78b4d7cba76e8a64d47c6d2c3874f3b251bc7053970adc8e35280537df85538

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page