Skip to content

WalePhenomenon/Hausa-NMT

Repository files navigation

HausaMT v1.0: Towards English–Hausa Neural Machine Translation

This is an ongoing work on Neural Machine Translation for English-Hausa. According to Sebastian Ruder, one of the biggest open problems for NLP is NMT for low-resource languages. NMT suffers a language diversity problem and growing up in a multi-lingual community with about 300 languages and thousands of dialects, I decided to work on NMT for the second largest Afro-Asiatic language after Arabic — Hausa Language. Hausa is also the third largest trade language across a larger swathe of West Africa after English and French. I have started working on this and the results are pretty good so far. I’m currently collaborating with scholars from the Niger-Volta Language Technologies Institute and working with some starter notebooks created by the Masakhane community.

Datasets and Summary

Pre-Processing and Training

  • We used Byte Pair Encoding (BPE) word-level tokenization
  • Trained using the Transformer Encoder-Decoder architecture on JoeyNMT
  • 30 epochs
  • Plateau scheduling
  • Learning rate: 0.0003
  • 4096 batch size
  • Xavier initializer (same used for embedding layer)
  • For transformer encoder and decoder: 6 layers, 4 heads, 256 embedding dim, 0.2 embedding dropout rate, 0.3 dropout rate, 256 hidden layer size

Model Files

  1. JW 300 - Transformer with BPE subword tokenization
  2. JW 300 - Transformer with word level tokenization
  3. All (JW300, Tanzil, Tatoeba & Wikimedia) - Transformer with BPE subword tokenization
  4. All (JW300, Tanzil, Tatoeba & Wikimedia) - Transformer with word level tokenization

Results

Sample Translation

Author: Adewale Akinfaderin (LinkedIn, Twitter)

About

Hausa-NMT: Empirical Study of Neural Machine translation for English-Hausa-English

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published