Skip to main content

Skipthoughts pretrained models for Pytorch

Project description

Skip-Thoughts.torch for Pytorcb

Skip-Thoughts.torch is a lightweight porting of skip-thought pretrained models from Theano to Pytorch.

Installation

  1. python3 with anaconda
  2. pytorch with/out CUDA

Install from pip

  1. pip install skipthoughts

Install from repo

  1. git clone https://github.com/Cadene/skip-thoughts.torch.git
  2. cd skip-thoughts.torch/pytorch
  3. python setup.py install

Available pretrained models

UniSkip

It uses the nn.GRU layer from torch with the cudnn backend. It is the fastest implementation, but the dropout is sampled after each time-step in the cudnn implementation... (equals bad regularization)

DropUniSkip

It uses the nn.GRUCell layer from torch with the cudnn backend. It is slightly slower than UniSkip, however the dropout is sampled once for all time-steps in a sequence (good regularization).

BayesianUniSkip

It uses a custom GRU layer with a torch backend. It is at least two times slower than UniSkip, however the dropout is sampled once for all time-steps for each Linear (best regularization).

BiSkip

Equivalent to UniSkip, but with a bi-sequential GRU.

Quick example

import torch
from torch.autograd import Variable
import sys
sys.path.append('skip-thoughts.torch/pytorch')
from skipthoughts import UniSkip

dir_st = 'data/skip-thoughts'
vocab = ['robots', 'are', 'very', 'cool', '<eos>', 'BiDiBu']
uniskip = UniSkip(dir_st, vocab)

input = Variable(torch.LongTensor([
    [1,2,3,4,0], # robots are very cool 0
    [6,2,3,4,5]  # bidibu are very cool <eos>
])) # <eos> token is optional
print(input.size()) # batch_size x seq_len

output_seq2vec = uniskip(input, lengths=[4,5])
print(output_seq2vec.size()) # batch_size x 2400

output_seq2seq = uniskip(input)
print(output_seq2seq.size()) # batch_size x seq_len x 2400

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

skipthoughts-0.0.1.tar.gz (9.8 kB view details)

Uploaded Source

Built Distribution

skipthoughts-0.0.1-py3-none-any.whl (9.1 kB view details)

Uploaded Python 3

File details

Details for the file skipthoughts-0.0.1.tar.gz.

File metadata

  • Download URL: skipthoughts-0.0.1.tar.gz
  • Upload date:
  • Size: 9.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.13.0 pkginfo/1.5.0.1 requests/2.22.0 setuptools/41.0.1 requests-toolbelt/0.9.1 tqdm/4.32.2 CPython/3.7.3

File hashes

Hashes for skipthoughts-0.0.1.tar.gz
Algorithm Hash digest
SHA256 8e4f87a29acd978db8a8b5f2d8ee0e9e1eef7d20845ce30c2dd3cf2f11b4f626
MD5 13b41cfb452a02809db0b700d92e9630
BLAKE2b-256 7b01b6c1126dc402e0d09d90f0fbaf1a6e95e4c4a77293d742d6300c68d164ca

See more details on using hashes here.

File details

Details for the file skipthoughts-0.0.1-py3-none-any.whl.

File metadata

  • Download URL: skipthoughts-0.0.1-py3-none-any.whl
  • Upload date:
  • Size: 9.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.13.0 pkginfo/1.5.0.1 requests/2.22.0 setuptools/41.0.1 requests-toolbelt/0.9.1 tqdm/4.32.2 CPython/3.7.3

File hashes

Hashes for skipthoughts-0.0.1-py3-none-any.whl
Algorithm Hash digest
SHA256 5c5664194d4d37a3d29698147bcd103c9cbb7573f698abad0248e813e6064a4e
MD5 04187fec0e63b409f815b4097dac63ff
BLAKE2b-256 3410e0b1f148f19b2a792f3811822feeb7f5561de5ae2fffe9b9544642447cdc

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page