• Stars
    star
    168
  • Rank 225,507 (Top 5 %)
  • Language
    Jupyter Notebook
  • License
    MIT License
  • Created almost 5 years ago
  • Updated 7 months ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Tools for shrinking fastText models (in gensim format)

Compress-fastText

This Python 3 package allows to compress fastText word embedding models (from the gensim package) by orders of magnitude, without significantly affecting their quality.

Here are some links to the models that have already been compressed.

This blogpost in Russian and this one in English give more details about the motivation and methods for compressing fastText models.

Note: gensim==4.0.0 has introduced some backward-incompatible changes:

  • With gensim<4.0.0, please use compress-fasttext<=0.0.7 (and optionally Russian models from our first release).
  • With gensim>=4.0.0, please use compress-fasttext>=0.1.0 (and optionally Russian or English models from our 0.1.0 release).
  • Some models are no longer supported in the new version of gensim+compress-fasttext (for example, multiple models from RusVectores that use compatible_hash=False).
  • For any particular model, compatibility should be determined experimentally. If you notice any strange behaviour, please report in the Github issues.

The package can be installed with pip:

pip install compress-fasttext[full]

If you are not going to perform matrix decomposition or quantization, you can install a variety with less dependencies:

pip install compress-fasttext

Model compression

You can use this package to compress your own fastText model (or one downloaded e.g. from RusVectores):

Compress a model in Gensim format:

import gensim
import compress_fasttext
big_model = gensim.models.fasttext.FastTextKeyedVectors.load('path-to-original-model')
small_model = compress_fasttext.prune_ft_freq(big_model, pq=True)
small_model.save('path-to-new-model')

Import a model in Facebook original format and compress it:

from gensim.models.fasttext import load_facebook_model
import compress_fasttext
big_model = load_facebook_model('path-to-original-model').wv
small_model = compress_fasttext.prune_ft_freq(big_model, pq=True)
small_model.save('path-to-new-model')

To perform this compression, you will need to pip install gensim==3.8.3 pqkmeans beforehand.

Different compression methods include:

  • matrix decomposition (svd_ft)
  • product quantization (quantize_ft)
  • optimization of feature hashing (prune_ft)
  • feature selection (prune_ft_freq)

The recommended approach is combination of feature selection and quantization (prune_ft_freq with pq=True).

Model usage

If you just need a tiny fastText model for Russian, you can download this 21-megabyte model. It's a compressed version of geowac_tokens_none_fasttextskipgram_300_5_2020 model from RusVectores.

If compress-fasttext is already installed, you can download and use this tiny model

import compress_fasttext
small_model = compress_fasttext.models.CompressedFastTextKeyedVectors.load(
    'https://github.com/avidale/compress-fasttext/releases/download/gensim-4-draft/geowac_tokens_sg_300_5_2020-100K-20K-100.bin'
)
print(small_model['спасибо'])
# [ 0.26762889  0.35489027 ...  -0.06149674] # a 300-dimensional vector
print(small_model.most_similar('котенок'))
# [('кот', 0.7391024827957153), ('пес', 0.7388300895690918), ('малыш', 0.7280327081680298), ... ]

The class CompressedFastTextKeyedVectors inherits from gensim.models.fasttext.FastTextKeyedVectors, but makes a few additional optimizations.

For English, you can use this tiny model, obtained by compressing the model by Facebook.

import compress_fasttext
small_model = compress_fasttext.models.CompressedFastTextKeyedVectors.load(
    'https://github.com/avidale/compress-fasttext/releases/download/v0.0.4/cc.en.300.compressed.bin'
)
print(small_model['hello'])
# [ 1.84736611e-01  6.32683930e-03  4.43901886e-03 ... -2.88431027e-02]  # a 300-dimensional vector
print(small_model.most_similar('Python'))
# [('PHP', 0.5252903699874878), ('.NET', 0.5027452707290649), ('Java', 0.4897131323814392),  ... ]

More compressed models for 101 various languages can be found at https://zenodo.org/record/4905385.

Example of application

In practical applications, you usually feed fastText embeddings to some other model. The class FastTextTransformer uses the scikit-learn interface and represents a text as the average of the embedding of its words. With it you can, for example, train a classifier on top of fastText to tell edible things from inedible ones:

import compress_fasttext
from sklearn.pipeline import make_pipeline
from sklearn.linear_model import LogisticRegression
from compress_fasttext.feature_extraction import FastTextTransformer

small_model = compress_fasttext.models.CompressedFastTextKeyedVectors.load(
    'https://github.com/avidale/compress-fasttext/releases/download/v0.0.4/cc.en.300.compressed.bin'
)

classifier = make_pipeline(
    FastTextTransformer(model=small_model), 
    LogisticRegression()
).fit(
    ['banana', 'soup', 'burger', 'car', 'tree', 'city'],
    [1, 1, 1, 0, 0, 0]
)
classifier.predict(['jet', 'train', 'cake', 'apple'])
# array([0, 0, 1, 1])

Notes

This code is heavily based on the navec package by Alexander Kukushkin and the blogpost by Andrey Vasnetsov about shrinking fastText embeddings.

More Repositories

1

encodechka

The tiniest sentence encoder for Russian language
Python
170
star
2

weirdMath

some examples of "high math" that we can play with
Jupyter Notebook
70
star
3

python-ruwordnet

A Python wrapper for the RuWordNet thesaurus
Jupyter Notebook
56
star
4

dialogic

Yet another common Python wrapper for Alice and Salut skills and bots in Telegram, VK, and Facebook
Python
25
star
5

constrained-linear-regression

Python
22
star
6

dependency-paraphraser

A sentence paraphraser based on dependency parsing and word embedding similarity.
Python
22
star
7

demo-alice-translate-skill

Пример навыка Алисы на Python с использованием интентов и состояния диалога
Python
10
star
8

tgalice

Yet another common Python wrapper for Alice skills and Telegram/Facebook/VK bots
Python
8
star
9

arxivarius

Python
8
star
10

mvideo_hack

Test problem for mvideo hackathon
Jupyter Notebook
7
star
11

sklearn_style

some python utils for machine learning
Python
6
star
12

example_telebot

Python
6
star
13

tgalice-yc-serverless-example

Python
3
star
14

botank

Библиотека для автоматического тестирования навыков Алисы и других чат-ботов
Python
3
star
15

rzd-russian-bots

Python
2
star
16

watchman_bot

Python
2
star
17

synonym-skill

Demo Alice skill with synonyms
Python
2
star
18

avidale

The README of my github profile
1
star
19

dale_competition

Data analysis competitions that I took part in
Jupyter Notebook
1
star
20

pushkin_rhyme

Python
1
star
21

alice-stretching

Python
1
star
22

displacy-demo

Python
1
star
23

word-mover-grammar

A constituency grammar parser with support of morphology and word embeddings
Python
1
star
24

dialogic-dashboard

a simple dashboard for viewing logs of dialogic apps
Python
1
star
25

DDTree

A few sklearn-based classifiers, mainly for purposes of credit scoring
Jupyter Notebook
1
star
26

cnop

Cross-nested ordered probit statistical models implementation in Stata
Rich Text Format
1
star