top of page

Deep Learning - Text Summarization

Goal of the project is to summarize the document as accurately as possible. This NLP task does not have a specific metric to describe accuracy of the model, hence metrics like Bleu Score and Rogue were used as an indicator.

Team: Pawanjeet Kaur, Shivam Duseja
Dataset: WikiHowAll.csv
Models: LSTM, LSTM with attention , GRU, GRU with attention, Stacked LSTM
Average time for one model to run on Colab Pro with corss validation : 4-5 hours

Best Bleu Score: 0.365

https://towardsdatascience.com/text-summarization-using-deep-neural-networks-e7ee7521d804#a87e-1aa55d239c9e

Why we chose Document Summarizer?

Increasing availability of online information has added more importance on the intensive research in the area of automatic text summarization within the Natural Language Processing (NLP) community. Text Summarization is basically an algorithm that shortens long pieces of sentences to produce a coherent summary that contains only the major points or the gist of the sentence which in turn saves time and helps us to identify the interest area. In this work, we build an abstractive text summarizer for different news articles using deep neural networks (Encoder-Decoder architecture).

For nearly the last half century, the field of summarization has been investigated by the NLP community. Earlier research works, proposed extraction of salient sentences from text using features like - phrase/word frequency (Luhn, 1958), position in the text (Baxendale, 1958) and key phrases (Edmundson, 1969).
Depending on the type of summary to be generated, different approaches of summarization have been proposed. The two major approaches for text summarization are - extractive and the abstractive technique. The extractive technique is mainly concerned with the extraction of words/phrases from the input sentence. On the other hand, the abstractive summary technique puts strong emphasis on the form - aiming to generate a grammatical summary thereby requiring advanced language modelling techniques.

Practical work after choosing the problem and dataset:
1. Data Exploration and Pre processing ( NLP task)
2. Exploratory Data Analysis - on sentence length and checking distribution of the data
3. Deep Model Design :
Seq2seq Model with Attention using LSTM ,
Seq2seq Model with Attention using GRU and Teacher Forcing
Three Stacked LSTM
4. Metric Calculation - Rouge , Bleu and F-Score

All these steps are described in detail in report attached. My teammate recently published his article on this project on toward data science.

Public Link: https://towardsdatascience.com/text-summarization-using-deep-neural-networks-e7ee7521d804#a87e-1aa55d239c9e

CONTACT ME

Pawanjeet Kaur

Software Engineer

​

Phone:

312-973-9407

 

Email:

sranpawan@gmail.com 

​

Thanks for contacting. I will be reaching out to you shortly :)

© 2020 By Pawanjeet Kaur. Proudly created with Wix.com

bottom of page