Skip to content

LittleDijkstraZ/PE_investigation

 
 

Repository files navigation

How does Transformers without PE Learn Positional Information?

Overview


This repository contains the code for our investigation of positional encoding in Transformers.

Important files

Notes on the implementation


This codebase forked from teaching arithmetic

  • Here contains our modifications.
  • NOPE has been implemented.
  • Control on Layerwise positional encoding has been added.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 97.0%
  • Python 2.3%
  • Shell 0.7%