Human in The Loop: Making the Most Out of Automated Question Answering Annotation #3176
-
Hello! In this article https://www.deepset.ai/blog/generate-questions-automatically-for-faster-annotation in the section "Human in The Loop: Making the Most Out of Automated Question Answering Annotation" there is a quote "You can then retrain your language model with the manually amended dataset, to make sure the model doesn't repeat the same mistakes." I want to first get a list of questions to my text using QuestionGenerationPipeline, after that, I correct this list of questions manually (1.leave valuable questions, 2.delete bad questions, 3.add other valuable questions) and add the resulting list of questions to the training dataset and then retrain the model for generating questions valhalla/t5-base-e2e-qg which is listed here https://haystack.deepset.ai/reference/question-generator What is the best way to train the model valhalla/t5-base-e2e-qg for the your Module question_generator using a list of questions received from the QuestionGenerationPipeline and corrected manually? Please tell me if I understood correctly: Does the squad_multitask dataset contain question-answer pairs? I have text. The model generated a set of questions for me. I manually corrected the received set of questions and want to retrain the valhalla/t5-small-e2e-qg model. I have the first list of questions that the model produced and the new list of questions that I got by manually improving the first list. How can I combine these two lists of questions (without answers) and use them as a dataset to retrain the valhalla/t5-small-e2e-qg model? I didn't find such example in your tutorial https://haystack.deepset.ai/tutorials/question-generation Please show an example code. Thanks a lot in advance. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 5 replies
-
Hey @SergiyBarskyy, If I understood you correctly, you want to adapt/fine-tune the valhalla/t5-small-e2e-qg question generation model on your data? |
Beta Was this translation helpful? Give feedback.
Hey @SergiyBarskyy,
If I understood you correctly, you want to adapt/fine-tune the valhalla/t5-small-e2e-qg question generation model on your data?
If so, the author of the model provided the details at https://github.com/patil-suraj/question_generation#fine-tuning
Fine-tuning such a model is out of scope for Haystack. But after you train it, then use it on Haystack 👍