Welcome to the Bahria University DSpace digital repository. DSpace is a digital service that collects, preserves, and distributes digital material. Repositories are important tools for preserving an organization's legacy; they facilitate digital preservation and scholarly communication.
dc.contributor.author | Engr. Muhammad Irfan, 01-242182-005 | |
dc.date.accessioned | 2020-11-17T05:51:49Z | |
dc.date.available | 2020-11-17T05:51:49Z | |
dc.date.issued | 2021 | |
dc.identifier.uri | http://hdl.handle.net/123456789/10302 | |
dc.description | Supervised by Dr. Shehzad Khalid | en_US |
dc.description.abstract | The reading of the long text is time consuming and sometimes understanding the context becomes difficult. Summaries are important specifically when we need to save our time and to understand the actual context of a long text corpus. Summarization is a technique to create a concise and accurate summary of a large script or a set of articles. In recent years abstractive text summarization tasks are most challenging in natural language processing. The existing encoder-decoder approaches have a potential issue. For the longer sequence of reviews, they need to compress all the necessary information into a fixed-length vector. This thesis aims to solve a very inherent task in data mining that is review summarization. Summary of the reviews has challenges that are dealing with variable length reviews, freestyle writing, and unstructured behavior. Our aim to create a shorter version of the review in abstractive manners while preserving the sentiment and points. In the decision-making process, it helps online customers to judge the product or service. To generate an optimal summary we have used a BRNN with LSTM's in the encoding layer. In the decoding layer, the attention mechanism is applied to the decoding cell that is just a two-layer LSTM with dropout. We have used Concept Net Number-Batch 3.0 word embeddings and Amazon Food reviews dataset. To reduce training loss and compute the learning rate of each parameter, we have used Adam Optimizer to reduce the loss function and for faster converge. We have achieved R1 38.75, R2 16.5, RL 36.25, and reduced the training loss with a new value of 0.031 for the whole dataset after removing the duplications. | en_US |
dc.language.iso | en | en_US |
dc.publisher | Electrical Engineering, Bahria University Engineering School Islamabad | en_US |
dc.relation.ispartofseries | MS(CE);T-0399 | |
dc.subject | Computer Engineering | en_US |
dc.title | Text mining (NLP) Abstractive text summarization using deep sequence models (T-0399) (MFN 8758) | en_US |
dc.type | MS Thesis | en_US |