Skip to content

Releases: Rishit-dagli/Fast-Transformer

Fast Transformer v0.2.0

16 Jan 09:43
8c7504e
Compare
Choose a tag to compare

✅ Bug Fixes / Improvements

  • Unit Tests for output rank and shape
  • Looser dependency requirements (now supports all TensorFlow versions >= 2.5.0)

Fast Transformer v0.1.0

03 Sep 05:34
5ed6f39
Compare
Choose a tag to compare

This is the initial release of Fast Transformer and implements Fast Transformer as a subclassed TensorFlow model.

Classes

  • FastAttention: Implements additive attention as a TensorFlow Keras layer, and supports using relative positional encodings.
  • PreNorm: Normalize the activations of the previous layer for each given example in a batch independently and apply some function to it, implemented as a TensorFlow Keras Layer.
  • FeedForward: Create a FeedForward neural net with two Dense layers and GELU activation, implemented as a TensorFlow Keras Layer.
  • FastTransformer: Implements the FastTransformer model using all the other classes, allows using rotary embeddings, weight tie projections, and converts to logits. Implemented as a TensorFlow Keras Model.