[D] BERT for seq2seq tasks
So am I right that BERT cannot currently be used for seq2seq tasks like machine translation or generating a response to an input sentence (like a general chatbot)?
If so, what are the best methods/architectures right now for seq2seq? Is bidirectional RNN /LSTM with attention still the best?
submitted by /u/AnonMLstudent
[link] [comments]