Skip to main content

A Simple Python Module for English Grapheme To Phoneme Conversion

Project description

[Update] * We removed TensorFlow from the dependencies. After all, it changes its APIs quite often, and we don’t expect you to have a GPU. Instead, NumPy is used for inference.

This module is designed to convert English graphemes (spelling) to phonemes (pronunciation). It is considered essential in several tasks such as speech synthesis. Unlike many languages like Spanish or German where pronunciation of a word can be inferred from its spelling, English words are often far from people’s expectations. Therefore, it will be the best idea to consult a dictionary if we want to know the pronunciation of some word. However, there are at least two tentative issues in this approach. First, you can’t disambiguate the pronunciation of homographs, words which have multiple pronunciations. (See a below.) Second, you can’t check if the word is not in the dictionary. (See b below.)

  • a. I refuse to collect the refuse around here. (rɪ|fju:z as verb vs. |refju:s as noun)

  • b. I am an activationist. (activationist: newly coined word which means n. A person who designs and implements programs of treatment or therapy that use recreation and activities to help people whose functional abilities are affected by illness or disability. from WORD SPY

For the first homograph issue, fortunately many homographs can be disambiguated using their part-of-speech, if not all. When it comes to the words not in the dictionary, however, we should make our best guess using our knowledge. In this project, we employ a deep learning seq2seq framework based on TensorFlow.

Algorithm

  1. Spells out arabic numbers and some currency symbols. (e.g. $200 -> two hundred dollars) (This is borrowed from Keith Ito’s code)

  2. Attempts to retrieve the correct pronunciation for homographs based on their POS)

  3. Looks up The CMU Pronouncing Dictionary for non-homographs.

  4. For OOVs, we predict their pronunciations using our neural net model.

Environment

  • python 3.x

Dependencies

  • numpy >= 1.13.1

  • nltk >= 3.2.4

  • python -m nltk.downloader “averaged_perceptron_tagger” “cmudict”

  • inflect >= 0.3.1

  • Distance >= 0.1.3

Installation

pip install g2p_en

OR

python setup.py install

nltk package will be automatically downloaded at your first run.

Training (Note that pretrained model is already included)

python train.py

Usage

from g2p_en import G2p

texts = ["I have $250 in my pocket.", # number -> spell-out
         "popular pets, e.g. cats and dogs", # e.g. -> for example
         "I refuse to collect the refuse around here.", # homograph
         "I'm an activationist."] # newly coined word
g2p = G2p()
for text in texts:
    out = g2p(text)
    print(out)
>>> ['AY1', ' ', 'HH', 'AE1', 'V', ' ', 'T', 'UW1', ' ', 'HH', 'AH1', 'N', 'D', 'R', 'AH0', 'D', ' ', 'F', 'IH1', 'F', 'T', 'IY0', ' ', 'D', 'AA1', 'L', 'ER0', 'Z', ' ', 'IH0', 'N', ' ', 'M', 'AY1', ' ', 'P', 'AA1', 'K', 'AH0', 'T', ' ', '.']
>>> ['P', 'AA1', 'P', 'Y', 'AH0', 'L', 'ER0', ' ', 'P', 'EH1', 'T', 'S', ' ', ',', ' ', 'F', 'AO1', 'R', ' ', 'IH0', 'G', 'Z', 'AE1', 'M', 'P', 'AH0', 'L', ' ', 'K', 'AE1', 'T', 'S', ' ', 'AH0', 'N', 'D', ' ', 'D', 'AA1', 'G', 'Z']
>>> ['AY1', ' ', 'R', 'IH0', 'F', 'Y', 'UW1', 'Z', ' ', 'T', 'UW1', ' ', 'K', 'AH0', 'L', 'EH1', 'K', 'T', ' ', 'DH', 'AH0', ' ', 'R', 'EH1', 'F', 'Y', 'UW2', 'Z', ' ', 'ER0', 'AW1', 'N', 'D', ' ', 'HH', 'IY1', 'R', ' ', '.']
>>> ['AY1', ' ', 'AH0', 'M', ' ', 'AE1', 'N', ' ', 'AE2', 'K', 'T', 'IH0', 'V', 'EY1', 'SH', 'AH0', 'N', 'IH0', 'S', 'T', ' ', '.']

May, 2018.

Kyubyong Park & Jongseok Kim

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

g2p_en-2.0.0.tar.gz (3.1 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

g2p_en-2.0.0-py3-none-any.whl (3.1 MB view details)

Uploaded Python 3

File details

Details for the file g2p_en-2.0.0.tar.gz.

File metadata

  • Download URL: g2p_en-2.0.0.tar.gz
  • Upload date:
  • Size: 3.1 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.13.0 pkginfo/1.5.0.1 requests/2.21.0 setuptools/41.0.0 requests-toolbelt/0.9.1 tqdm/4.31.1 CPython/3.6.8

File hashes

Hashes for g2p_en-2.0.0.tar.gz
Algorithm Hash digest
SHA256 678d6bf36ce2c14d360bdf1a0ca0355d22a299a981b4b16917125cbd2914ac6c
MD5 814e68be3819508afe9103188385d8fc
BLAKE2b-256 d98e236d430e8440ae5ffaf32843ac99cd1257b1337cc9a8d1b99db76b54c7fd

See more details on using hashes here.

File details

Details for the file g2p_en-2.0.0-py3-none-any.whl.

File metadata

  • Download URL: g2p_en-2.0.0-py3-none-any.whl
  • Upload date:
  • Size: 3.1 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.13.0 pkginfo/1.5.0.1 requests/2.21.0 setuptools/41.0.0 requests-toolbelt/0.9.1 tqdm/4.31.1 CPython/3.6.8

File hashes

Hashes for g2p_en-2.0.0-py3-none-any.whl
Algorithm Hash digest
SHA256 ebecc57f1033dcc296f4b184cc41c261c24be0ff006d98045bbdcf33562e6616
MD5 85f319f2701a58140f4f92b966bec4f3
BLAKE2b-256 6ec0328d4f4a40e6d2cd699af7987c2a4402d98998e4650982a6efe9eb52f5fe

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page