Skip to main content

A collection of syntactic metrics to calculate (dis)similarities between source and target sentences.

Project description

Example notebooks

A couple example notebooks exist, each with a different grade of automation for the initialisation of the aligned object. Once an aligned object has been created, the functionality is identical.

  • High automation: automate all the things. Tokenisation, parsing, and word alignment is done automatically [Try on Colab]
  • Normal automation: the typical scenario where you have tokenised and aligned text that is not parsed yet [Try on Colab]
  • No automation: full-manual mode, where you provide all the required information, including dependency labels and heads [Try on Colab]
  • Monolingual: in this example we rely on spaCy to compare two English sentences and calculate semantic similarity between aligned words [Try on Colab]

Installation

Requires Python 3.7 or higher. To keep the overhead low, a default parser is NOT installed. Currently both spaCy and stanza are supported and you can choose which one to use. Stanza is recommended for bilingual research (because it is ensured that all of its models use Universal Dependencies), but spaCy can be used as well. The latter is especially used for monolingual comparisons, or if you are not interested in the linguistic comparisons and only require word reordering metrics.

A pre-release is available on PyPi. You can install it with pip as follows.

# Install with stanza (recommended)
pip install astred[stanza]
# ... or install with spacy
pip install astred[spacy]
# ... or install with both and decide later
pip install astred[parsers]

If you want to use spaCy, you have to make sure that you install the required models manually, which cannot be automated.

Automatic Word Alignment

Automatic word alignment is supported by using a modified version of Awesome Align under the hood. This is a neural word aligner that uses transfer learning with multilingual models to do word alignment. It does require some manual installation work. Specifically, you need to install the astred_compat branch from this fork. If you are using pip, you can run the following command:

pip install git+https://github.com/BramVanroy/awesome-align.git@astred_compat

Awesome Align requires PyTorch, like stanza above.

If it is installed, you can initialize AlignedSentences without providing word alignments. Those will be added automatically behind the scenes. See this example notebook [Try on Colab] for more.

sent_en = Sentence.from_text("I like eating cookies", "en")
sent_nl = Sentence.from_text("Ik eet graag koekjes", "nl")

# Word alignments do not need to be added on init:
aligned = AlignedSentences(sent_en, sent_nl)

Keep in mind however that automatic alignment will never have the same quality as manual alignments. Use with caution! I highly suggest reading the paper of Awesome Align to see whether it is a good pick for you.

License

Licensed under Apache License Version 2.0. See the LICENSE file attached to this repository.

Citation

Please cite our papers if you use this library.

Vanroy, B., De Clercq, O., Tezcan, A., Daems, J., & Macken, L. (2021). Metrics ofsyntactic equivalence to assess translation difficulty. In M. Carl (Ed.), Explorations in empirical translation process research (Vol. 3, pp. 259–294). Cham, Switzerland: Springer International Publishing. https://doi.org/10.1007/978-3-030-69777-8_10

@incollection{vanroy2021metrics,
    title = {Metrics of syntactic equivalence to assess translation difficulty},
    booktitle = {Explorations in empirical translation process research},
    author = {Vanroy, Bram and De Clercq, Orph{\'e}e and Tezcan, Arda and Daems, Joke and Macken, Lieve},
    editor = {Carl, Michael},
    year = {2021},
    series = {Machine {{Translation}}: {{Technologies}} and {{Applications}}},
    volume = {3},
    pages = {259--294},
    publisher = {{Springer International Publishing}},
    address = {{Cham, Switzerland}},
    isbn = {978-3-030-69776-1},
    url = {https://link.springer.com/chapter/10.1007/978-3-030-69777-8_10},
    doi = {10.1007/978-3-030-69777-8_10}
}

Vanroy, B., Schaeffer, M., & Macken, L. (2021). Comparing the Effect of Product-Based Metrics on the Translation Process. Frontiers in Psychology, 12. https://doi.org/10.3389/fpsyg.2021.681945

@article{vanroy2021comparing,
    publisher = {Frontiers},
    author = {Vanroy, Bram and Schaeffer, Moritz and Macken, Lieve},
    title = {Comparing the effect of product-based metrics on the translation process},
    year = {2021},
    journal = {Frontiers in Psychology},
    volume = {12},
    issn = {1664-1078},
    url = {https://www.frontiersin.org/article/10.3389/fpsyg.2021.681945},
    doi = {10.3389/fpsyg.2021.681945},
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Files for astred, version 0.9.6
Filename, size File type Python version Upload date Hashes
Filename, size astred-0.9.6-py3-none-any.whl (29.2 kB) File type Wheel Python version py3 Upload date Hashes View
Filename, size astred-0.9.6.tar.gz (28.7 kB) File type Source Python version None Upload date Hashes View

Supported by

AWS AWS Cloud computing Datadog Datadog Monitoring DigiCert DigiCert EV certificate Facebook / Instagram Facebook / Instagram PSF Sponsor Fastly Fastly CDN Google Google Object Storage and Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Salesforce Salesforce PSF Sponsor Sentry Sentry Error logging StatusPage StatusPage Status page