Skip to content

riju-stone/chatbot

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

47 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Chatbots

References

Recurrent Neural Networks

In order to understand the concepts of Seq2Seq models or Transformer models, the first thing to know is about RNNs,

Sequence to Sequence Model

A Sequence to Sequence Model aims to map a fixed-length input with a fixed-length output where the length of the input and output may differ.

Transformer Model (Seq2Seq with Attention)

The Transformer in NLP is a novel architecture that aims to solve sequence-to-sequence tasks while handling long-range dependencies with ease.

“The Transformer is the first transduction model relying entirely on self-attention to compute representations of its input and output without using sequence-aligned RNNs or convolution.”

Further Resources