The GPT-2 model released by huggingface, as part of the set of pre-trained models available on Pytorch Hub, has been fine tuned using a dataset of news extracted from Huffpost web.
The main goal of this project is to explore how models can output texts resembling the structure and style of news from a traditional newspaper. For that purpose, we've developed a simple web showing the latest news created by the trained model.
The model (1,4 GB) is not hosted in this repository as it would exceed the max limit allowed by Github. But the file training_medium_news.ipynb
it's the Jupyter Notebook that we used for training the model, you can either use that script to train it yourself or download it from this Google Drive shared URL.