Skip to content

Open-source Python toolkit focused on deep learning with ordinal methodologies

License

Notifications You must be signed in to change notification settings

ayrna/dlordinal

Repository files navigation

Welcome to dlordinal

dlordinal is a Python library that unifies many recent deep ordinal classification methodologies available in the literature. Developed using PyTorch as underlying framework, it implements the top performing state-of-the-art deep learning techniques for ordinal classification problems. Ordinal approaches are designed to leverage the ordering information present in the target variable. Specifically, it includes loss functions, various output layers, dropout techniques, soft labelling methodologies, and other classification strategies, all of which are appropriately designed to incorporate the ordinal information. Furthermore, as the performance metrics to assess novel proposals in ordinal classification depend on the distance between target and predicted classes in the ordinal scale, suitable ordinal evaluation metrics are also included.

The latest dlordinal release is v2.2.0.

Overview
CI/CD !codecov !docs !python
Code ![pypi] ![binder] !black Linter: Ruff

Table of Contents

⚙️ Installation

dlordinal v2.2.0 is the last version, supported by Python 3.8, Python 3.9 and Python 3.10.

The easiest way to install dlordinal is via pip:

pip install dlordinal

🚀 Getting started

The best place to get started with dlordinal is the tutorials directory.

Below we provide a quick example of how to use some elements of dlordinal, such as a dataset, a loss function or some metrics.

Loading an ordinal benchmark dataset

The FGNet is a well-known benchmark dataset that is commonly used to benchmark ordinal classification methodologies. The dataset is composed of facial images and is labelled with different age categories. It can be downloaded and loaded into Python by simply using the dlordinal.datasets.FGNet class.

import numpy as np
from dlordinal.datasets import FGNet
from torchvision.transforms import Compose, ToTensor

fgnet_train = FGNet(
    root="./datasets",
    train=True,
    target_transform=np.array,
    transform=Compose([ToTensor()]),
)
fgnet_test = FGNet(
    root="./datasets",
    train=False,
    target_transform=np.array,
    transform=Compose([ToTensor()]),
)

Training a CNN model using the skorch library

This example shows how to train a CNN model using the NeuralNetClassifier from the skorch library and the TriangularCrossEntropy from dlordinal as optimisation criterion.

import numpy as np
from dlordinal.datasets import FGNet
from dlordinal.losses import TriangularCrossEntropyLoss
from dlordinal.metrics import amae, mmae
from skorch import NeuralNetClassifier
from torch import nn
from torch.optim import Adam
from torchvision import models
from torchvision.transforms import Compose, ToTensor

# Download the FGNet dataset
fgnet_train = FGNet(
    root="./datasets",
    train=True,
    target_transform=np.array,
    transform=Compose([ToTensor()]),
)
fgnet_test = FGNet(
    root="./datasets",
    train=False,
    target_transform=np.array,
    transform=Compose([ToTensor()]),
)

num_classes_fgnet = len(fgnet_train.classes)

# Model
model = models.resnet18(weights="IMAGENET1K_V1")
model.fc = nn.Linear(model.fc.in_features, num_classes_fgnet)

# Loss function
loss_fn = TriangularCrossEntropyLoss(num_classes=num_classes_fgnet)

# Skorch estimator
estimator = NeuralNetClassifier(
    module=model,
    criterion=loss_fn,
    optimizer=Adam,
    lr=1e-3,
    max_epochs=25,
)

estimator.fit(X=fgnet_train, y=fgnet_train.targets)
train_probs = estimator.predict_proba(fgnet_train)
test_probs = estimator.predict_proba(fgnet_test)

# Metrics
amae_metric = amae(np.array(fgnet_test.targets), test_probs)
mmae_metric = mmae(np.array(fgnet_test.targets), test_probs)
print(f"Test AMAE: {amae_metric}, Test MMAE: {mmae_metric}")

📖 Documentation

Sphinx is a documentation generator tool that is commonly used in the Python ecosystem. It allows developers to write documentation in a markup language called reStructuredText (reST) and generates HTML, PDF, and other formats from it. Sphinx provides a powerful and flexible way to document code, making it easier for developers to create comprehensive and user-friendly documentation for their projects.

To document dlordinal, it is necessary to install all documentation dependencies:

pip install -e '.[docs]'

Then access the docs/ directory:

docs/
↳ api.rst
↳ conf.py
↳ distributions.rst
↳ references.bib
↳ ...

If a new module is created in the software project, the api.rst file must be modified to include the name of the new module:

.. _api:

=============
API Reference
=============

This is the API for the **dlordinal** package.

.. toctree::
   :maxdepth: 2
   :caption: Contents:

   datasets
   dropout
   output_layers
   losses
   metrics
   wrappers
   soft_labelling
   ***NEW_MODULE***

Afterwards, a new file in .rst format associated to the new module must be created, specifying the automatic inclusion of documentation from the module files containing a docstring, and the inclusion of the bibliography if it exists within any of them.

docs/
↳ api.rst
↳ conf.py
↳ distributions.rst
↳ new_module.rst
↳ references.bib
↳ ...
.. _new_module:

New Module
==========

.. automodule:: dlordinal.new_module
    :members:

.. footbibliography::

Finally, if any new bibliographic citations have been added, they should be included in the references.bib file.

Collaborating

Code contributions to the dlordinal project are welcomed via pull requests. Please, contact the maintainers (maybe opening an issue) before doing any work to make sure that your contributions align with the project.

Guidelines for code contributions

  • You can clone the repository and then install the library from the local repository folder:
git clone git@github.com:ayrna/dlordinal.git
pip install ./dlordinal
  • In order to set up the environment for development, install the project in editable mode and include the optional dev requirements:
pip install -e '.[dev]'
  • Install the pre-commit hooks before starting to make any modifications:
pre-commit install
  • Write code that is compatible with all supported versions of Python listed in the pyproject.toml file.
  • Create tests that cover the common cases and the corner cases of the code.
  • Preserve backwards-compatibility whenever possible, and make clear if something must change.
  • Document any portions of the code that might be less clear to others, especially to new developers.
  • Write API documentation as docstrings.