Generative adversarial imitation learning to produce a proxy for the reward function present in dialogue.
-
Updated
Dec 25, 2020 - Jupyter Notebook
Generative adversarial imitation learning to produce a proxy for the reward function present in dialogue.
A Seq2Seq Attention chatbot deployed on Heroku
Neural Machine Translation by Seq2Seq Model with Attention layer
Seq2Seq model that restores punctuation on English input text.
Seq2seq-attention house price prediction.
This repository is base on Pytorch Tutorial with some experiments and refined.
Neural Machine Translation using LSTMs and Attention mechanism. Two approaches were implemented, models, one without out attention using repeat vector, and the other using encoder decoder architecture and attention mechanism.
A few approaches using sequence to sequence (seq2seq) architecture to solve semantice parsing problem
Seq2Seq Neural Machine Translation Task, Completed as a part of CS462-NLP Coursework
Some natural language processing networks from scratch in PyTorch for personal educational purposes.
Sequence to sequence learning for GEC task using several deep models.
I replicate and make the original Seq2Seq from PyTorch tutorials to be easy to use and adapt.
Experimentation of converting English to Pig Latin via a variety of Vanilla-seq2seq networks, Attention-mechanism based models and Transformer based Machine translation system.
French to English neural machine translation trained on multi30k dataset.
基于Seq2Seq+Attention模型的Textsum文本自动摘要
This repository contains the code for a speech to speech translation system created from scratch for digits translation from English to Tamil
Sequence-to-sequence model implementations including RNN, CNN, Attention, and Transformers using PyTorch
Chatbot using Seq2Seq model using Tensorflow
Generates summary of a given news article. Used attention seq2seq encoder decoder model.
load point forecast
Add a description, image, and links to the seq2seq-attn topic page so that developers can more easily learn about it.
To associate your repository with the seq2seq-attn topic, visit your repo's landing page and select "manage topics."