Quantized Approximately Orthogonal Recurrent Neural Networks - IRT Saint Exupéry - Institut de Recherche Technologique Accéder directement au contenu
Pré-Publication, Document De Travail Année : 2024

Quantized Approximately Orthogonal Recurrent Neural Networks

Résumé

In recent years, Orthogonal Recurrent Neural Networks (ORNNs) have gained popularity due to their ability to manage tasks involving long-term dependencies, such as the copy-task, and their linear complexity. However, existing ORNNs utilize full precision weights and activations, which prevents their deployment on compact devices. In this paper, we explore the quantization of the weight matrices in ORNNs, leading to Quantized approximately Orthogonal RNNs (QORNNs). The construction of such networks remained an open problem, acknowledged for its inherent instability. We propose and investigate two strategies to learn QORNN by combining quantization-aware training (QAT) and orthogonal projections. We also study post-training quantization of the activations for pure integer computation of the recurrent loop. The most efficient models achieve results similar to state-of-the-art full-precision ORNN, LSTM and FastRNN on a variety of standard benchmarks, even with 4-bits quantization.
Fichier principal
Vignette du fichier
qornn_neurips.pdf (675.75 Ko) Télécharger le fichier
Origine Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-04434011 , version 1 (02-02-2024)
hal-04434011 , version 2 (07-06-2024)

Identifiants

Citer

Armand Foucault, Franck Mamalet, François Malgouyres. Quantized Approximately Orthogonal Recurrent Neural Networks. 2024. ⟨hal-04434011v2⟩
40 Consultations
26 Téléchargements

Altmetric

Partager

Gmail Mastodon Facebook X LinkedIn More