Can BERT be finetuned for text generation? #1037
-
Maybe someone can clarify this for me or point me in the right direction. So I'm aware that you can finetune the pretrained GPT2 model to generate text for a more specialized case, I'm clear on that. I've seen a lot of examples where BERT is being finetuned, but I'm still wondering: can BERT be finetuned for text generation? Or is there some other more lightweight pretrained model which could be finetuned for text generation, except BERT or GPT2? Preferably a transformer-style model, I'm also aware of using LSTM or RNN models (with GRU) to do text generation. |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments
-
you can't use an encoder model for text generation. and bert is an encoder model. only decoder or encoder-decoder models can be used for text generation. |
Beta Was this translation helpful? Give feedback.
-
#1030 (comment) |
Beta Was this translation helpful? Give feedback.
#1030 (comment)
use this solution with training GPT from scratch example it worked.