/recommendation-systems-final-project

Exploring Bloom embeddings as a compression technique for recommendation algorithms. Aimed at reducing the size of large input and output dimensionalities to enhance training and deployment efficiency on devices with limited hardware. This project evaluates Bloom embeddings using various hash functions and compares them with alternative methods.

Primary LanguagePython

Recommendation Systems - Final Project

Getting Started

Recommendation algorithms tend to have large input and output dimensionalities that dominate their overall size. This makes them difficult to train, due to the limited memory of graphical processing units, and difficult to deploy on mobile devices with limited hardware. To address these difficulties, we propose Bloom embeddings, a compression technique that can be applied to the input and output of neural network models dealing with sparse high-dimensional binary-coded instances.

Motivation

We would like to evaluate Bloom embeddings with different hash functions and compare them against these alternative methods.

Bloom embedding layers

Large embedding layers are a performance problem for fitting models. Although the gradients are sparse, PyTorch updates the entire embedding layer at every backward pass. Computation time is wasted on applying zero gradient steps to the whole embedding matrix.

To alleviate this problem, we can use a smaller underlying embedding layer, and probabilistically hash users and items into that smaller space. With good hash functions, collisions should be rare, and we should observe fitting speedups without a decreased inaccuracy.

Advantages of the Bloom embeddings Layers:

  • They are computationally efficient and do not seriously compromise the accuracy of the model.
  • They do not require any change to the core model architecture or training configuration.
  • "On-the-fly" constant-time operation.
  • Zero or marginal space requirements.
  • Training time speedups.

Implementation

This implementation uses embedding layers indexed by hashed indices. Vectors retrieved by each hash function are averaged together to yield the final embedding.
For hashing, we used different hash functions such as: MurmurHash, xxHash, MD5, SHA1, SHA256, MetroHash & FarmHash, and we performed hashing on the indices with a different seed, modulo the size of the compressed embedding layer. The hash mapping is computed once at the start of training and indexed into for every minibatch. To exploit this sparsity structure, we simply skip zero entries in the input vector.


Installations

echo "source $HOME/miniconda3/etc/profile.d/conda.sh" >> ~/.bashrc
  • Create and activate your virtual environment:
conda env update -f environment.yml      # create the virtual environemnt
conda env list                           # make sure the environemnt was added
conda activate recommandation-systems    # activate it

Running Example

from datasets.movielens import get_movielens_dataset
from evaluations.cross_validation import random_train_test_split
from evaluations.evaluation import mrr_score
from factorization.implicit import ImplicitFactorizationModel
from model_utils.layers import BloomEmbedding
from model_utils.networks import Net

dataset = get_movielens_dataset(variant='100K')
train, test = random_train_test_split(dataset)

for hash_function in ('MurmurHash', 'xxHash', 'MD5', 'SHA1', 'SHA256'):
    user_embeddings = BloomEmbedding(dataset.num_users, 32, compression_ratio=0.4, num_hash_functions=2, hash_function=hash_function)
    item_embeddings = BloomEmbedding(dataset.num_items, 32, compression_ratio=0.4, num_hash_functions=2, hash_function=hash_function)

    network = Net(dataset.num_users, dataset.num_items, user_embedding_layer=user_embeddings, item_embedding_layer=item_embeddings)
    model = ImplicitFactorizationModel(n_iter=1, loss='bpr', batch_size=1024, learning_rate=1e-2, l2=1e-6, representation=network, use_cuda=False)
    model.fit(train, verbose=True)

    mrr = mrr_score(model, test)
    print(f'MRR Score: {mrr.mean()}')

Datasets

References

Authors

  • Sharon Mordechai.
  • Eran Krakovsky.