shoelace
shoelace copied to clipboard
Neural Learning to Rank using Chainer
:warning: This library has been deprecated and archived in favor of https://github.com/rjagerman/pytorchltr.
Shoelace is a neural Learning to Rank library using Chainer. The goal is to make it easy to do offline learning to rank experiments on annotated learning to rank data.
Documentation
Comprehensive documentation is available online here
Requirements
python3
numpy >= 1.12.0
chainer >= 2.0.0
Installation
pip install shoelace
Features
Dataset loading facilities
We currently provide ability to load learning to rank datasets (SVMRank format) into chainer.
from shoelace.dataset import LtrDataset
with open('./dataset.txt', 'r') as file:
dataset = LtrDataset.load_txt(file)
Additionally, we provide minibatch iterators for Learning to Rank datasets. These generate variable-sized minibatches, where each minibatch represents one query and all associated query-document instances. You can additionally specify whether the iterator should repeat infinitely and/or shuffle the data on every epoch.
from shoelace.iterator import LtrIterator
iterator = LtrIterator(dataset, repeat=True, shuffle=True)
Loss functions
Currently we provide implementations for the following loss functions
- Top-1 ListNet:
shoelace.loss.listwise.listnet
- ListMLE:
shoelace.loss.listwise.listmle
- ListPL:
shoelace.loss.listwise.listpl
Example
Here is an example script that will train up a single-layer linear neural network with a ListNet loss function:
from shoelace.dataset import LtrDataset
from shoelace.iterator import LtrIterator
from shoelace.loss.listwise import listnet
from chainer import training, optimizers, links, Chain
from chainer.training import extensions
# Load data and set up iterator
with open('./path/to/ranksvm.txt', 'r') as f:
training_set = LtrDataset.load_txt(f)
training_iterator = LtrIterator(training_set, repeat=True, shuffle=True)
# Create neural network with chainer and apply loss function
predictor = links.Linear(None, 1)
class Ranker(Chain):
def __call__(self, x, t):
return listnet(self.predictor(x), t)
loss = Ranker(predictor=predictor)
# Build optimizer, updater and trainer
optimizer = optimizers.Adam()
optimizer.setup(loss)
updater = training.StandardUpdater(training_iterator, optimizer)
trainer = training.Trainer(updater, (40, 'epoch'))
trainer.extend(extensions.ProgressBar())
# Train neural network
trainer.run()
Citing
If you find this project useful in your research, please cite the following paper in your publication(s):
Rolf Jagerman, Julia Kiseleva and Maarten de Rijke. "Modeling Label Ambiguity for Neural List-Wise Learning to Rank" (2017)
@article{jagerman2017modeling,
title={Modeling Label Ambiguity for Neural List-Wise Learning to Rank},
author={Jagerman, Rolf and Kiseleva, Julia and de Rijke, Maarten},
journal={arXiv preprint arXiv:1707.07493},
year={2017}
}