generated_from_trainer

BART-base Question Generation

This model is a fine-tuned version of facebook/bart-base on different questions and answering dataset. It was trained to generation question using two different approaches, <b> Casual-Generation </b> and <b> Context-based-Generation </b>.

Model description

The model takes context as an input sequence, and will generate a full question sentence as an output sequence. There are two ways the model can be queried produce the questions:

The input sequence can then be encoded and passed as the input_ids argument in the model's generate() method.

limitations

The model was trained on only a limited amount of data hence questions might be poor quality. In addition the questions generated have style similar to that of the training data.

Training and evaluation data

The dataset used to train the model comprises the training datasets from:

After preprocessing the data from the above listed datasets, we had 408372 examples for training the model and 25k for development and 18k for testing.

Training procedure

The model is trained (finetuned) for 5 epochs with the hyperparameters listed below:

Training hyperparameters

The following hyperparameters were used during training:

At the end of 5 epochs, the Evaluation loss was: 1.64 and the training loss was: 0.9671.

Framework versions