Seq2Seq Architecture based DL chatbot
Published:
This is a Proof-of-Concept Project on Seq2Seq Neural Architecture.
- Trained on Movie-lens dataset having conversations between different people by building a Seq2Seq neural architecture.
- It is trained on 100 epochs with a batch size of 64, and number of layers =3. The encoding and decoding embedding size are of 512 each.
- Dropout regularization used to remove any kind of overfitting and the losses are optimized using Adam optimizer.
- All the sentences in a batch, whether they are questions/ answers must have the same length. Hence ‘PAD’ tokens used.