summarization

BART for Gigaword

from transformers import AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("a1noack/bart-large-gigaword")

Summary generation

input_ids_list = tokenizer(text_list, truncation=True, max_length=128, 
       return_tensors='pt', padding=True)['input_ids']
output_ids_list = model.generate(input_ids_list, min_length=0)
outputs_list = tokenizer.batch_decode(output_ids_list, skip_special_tokens=True, 
       clean_up_tokenization_spaces=False)