Skip to main content

BERT token level embedding with MxNet

Project description

Bert Embeddings

Build Status PyPI version

BERT, published by Google, is new way to obtain pre-trained language model word representation. Many NLP tasks are benefit from BERT to get the SOTA.

The goal of this project is to obtain the sentence and token embedding from BERT's pre-trained model. In this way, instead of building and do fine-tuning for an end-to-end NLP model, you can build your model by just utilizing the sentence or token embedding.

This project is implemented with @MXNet. Special thanks to @gluon-nlp team.

Install

pip install bert-embedding
pip install https://github.com/dmlc/gluon-nlp/tarball/master
# If you want to run on GPU machine, please install `mxnet-cu92`.
pip install mxnet-cu92

This project use API from gluonnlp==0.5.1, which hasn't been released yet. Once 0.5.1 is release, it is not necessary to install gluonnlp from source.

Usage

from bert_embedding import BertEmbedding

bert_abstract = """We introduce a new language representation model called BERT, which stands for Bidirectional Encoder Representations from Transformers.
 Unlike recent language representation models, BERT is designed to pre-train deep bidirectional representations by jointly conditioning on both left and right context in all layers.
 As a result, the pre-trained BERT representations can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks, such as question answering and language inference, without substantial task-specific architecture modifications. 
BERT is conceptually simple and empirically powerful. 
It obtains new state-of-the-art results on eleven natural language processing tasks, including pushing the GLUE benchmark to 80.4% (7.6% absolute improvement), MultiNLI accuracy to 86.7 (5.6% absolute improvement) and the SQuAD v1.1 question answering Test F1 to 93.2 (1.5% absolute improvement), outperforming human performance by 2.0%."""
sentences = bert_abstract.split('\n')
bert = BertEmbedding()
result = bert.embedding(sentences)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

bert_embedding-0.1.1.dev1549820793.tar.gz (5.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

bert_embedding-0.1.1.dev1549820793-py3-none-any.whl (10.2 kB view details)

Uploaded Python 3

File details

Details for the file bert_embedding-0.1.1.dev1549820793.tar.gz.

File metadata

  • Download URL: bert_embedding-0.1.1.dev1549820793.tar.gz
  • Upload date:
  • Size: 5.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.12.1 pkginfo/1.5.0.1 requests/2.18.4 setuptools/40.8.0 requests-toolbelt/0.9.1 tqdm/4.31.0 CPython/3.5.6

File hashes

Hashes for bert_embedding-0.1.1.dev1549820793.tar.gz
Algorithm Hash digest
SHA256 081dad120bfa221b134f44a691b1d91338f8c05f2371df8b2543902c2136f66b
MD5 f5026aa48f5aa746107e22a4625120f3
BLAKE2b-256 365a7652259e0c7409236889db496221136224dcc8c63a3285fbf01a137169f3

See more details on using hashes here.

File details

Details for the file bert_embedding-0.1.1.dev1549820793-py3-none-any.whl.

File metadata

  • Download URL: bert_embedding-0.1.1.dev1549820793-py3-none-any.whl
  • Upload date:
  • Size: 10.2 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.12.1 pkginfo/1.5.0.1 requests/2.18.4 setuptools/40.8.0 requests-toolbelt/0.9.1 tqdm/4.31.0 CPython/3.5.6

File hashes

Hashes for bert_embedding-0.1.1.dev1549820793-py3-none-any.whl
Algorithm Hash digest
SHA256 9cdcc2daf913b577b0571237df189a66b9776811cd417c50ec3b2ae9798a244f
MD5 0a41c38d42631d439cadc34aaafd389a
BLAKE2b-256 35a0a863f9160b5820b443ea108c98bdc5a1470dff3a21257c6292e283866ed8

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page