Text mining (NLP) Abstractive text summarization using deep sequence models (T-0399) (MFN 8758)

Welcome to DSpace BU Repository

Welcome to the Bahria University DSpace digital repository. DSpace is a digital service that collects, preserves, and distributes digital material. Repositories are important tools for preserving an organization's legacy; they facilitate digital preservation and scholarly communication.

Show simple item record

dc.contributor.author Engr. Muhammad Irfan, 01-242182-005
dc.date.accessioned 2020-11-17T05:51:49Z
dc.date.available 2020-11-17T05:51:49Z
dc.date.issued 2021
dc.identifier.uri http://hdl.handle.net/123456789/10302
dc.description Supervised by Dr. Shehzad Khalid en_US
dc.description.abstract The reading of the long text is time consuming and sometimes understanding the context becomes difficult. Summaries are important specifically when we need to save our time and to understand the actual context of a long text corpus. Summarization is a technique to create a concise and accurate summary of a large script or a set of articles. In recent years abstractive text summarization tasks are most challenging in natural language processing. The existing encoder-decoder approaches have a potential issue. For the longer sequence of reviews, they need to compress all the necessary information into a fixed-length vector. This thesis aims to solve a very inherent task in data mining that is review summarization. Summary of the reviews has challenges that are dealing with variable length reviews, freestyle writing, and unstructured behavior. Our aim to create a shorter version of the review in abstractive manners while preserving the sentiment and points. In the decision-making process, it helps online customers to judge the product or service. To generate an optimal summary we have used a BRNN with LSTM's in the encoding layer. In the decoding layer, the attention mechanism is applied to the decoding cell that is just a two-layer LSTM with dropout. We have used Concept Net Number-Batch 3.0 word embeddings and Amazon Food reviews dataset. To reduce training loss and compute the learning rate of each parameter, we have used Adam Optimizer to reduce the loss function and for faster converge. We have achieved R1 38.75, R2 16.5, RL 36.25, and reduced the training loss with a new value of 0.031 for the whole dataset after removing the duplications. en_US
dc.language.iso en en_US
dc.publisher Electrical Engineering, Bahria University Engineering School Islamabad en_US
dc.relation.ispartofseries MS(CE);T-0399
dc.subject Computer Engineering en_US
dc.title Text mining (NLP) Abstractive text summarization using deep sequence models (T-0399) (MFN 8758) en_US
dc.type MS Thesis en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search DSpace


Advanced Search

Browse

My Account