Skip to content

This repo contains the experiemensts which i did as a part of my research work on Text Summarization

License

Notifications You must be signed in to change notification settings

Darshan2104/Text-Summarization-experiments

Repository files navigation

Text-Summarization-experiments On News Data

This repo contains the experiments that I did as a part of my research work on Abstractive Text Summarization.

  • Hear is a script I used for training.
  • Here is a detailed report of the whole experiment.

Tech stack:

* Huggingface
* Adapter
* Datasets

Experiments:

1. PEGASUS model is used for Abstractive summarization. This model was pre-trained with semi-supervised learning techniques like Gap Sentence Generation(GSG) and Mask Language modeling (MLM).
2. We fine-tuned it on the news dataset for the downstream task- Abstractive text summarization.
3. For only 100 records and that was also with a base model, it gave us a ROUGE score of ~21 (state of the art ~46 with the entire dataset).
  • This was a part of my final year project here presentation.
1. We fine-tuned T5-small and T5-base model on the supercomputer (param shavak).
2. Also we Trained Adapter for the same model and with the same dataset.
3. We also experimented summary to headline and text to headline generation.

- When both results were compared, it was almost similar, the thing was, for base model fine-tuning took around 22 days, and compared to the adapter took only 6 days!!
- text to headline was better compared to summary to headline.

About

This repo contains the experiemensts which i did as a part of my research work on Text Summarization

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published