MeetSum: Transforming Meeting Transcript Summarization using Transformers! (original) (raw)

Assessing the Efficacy of LSTM, Transformer, and RNN Architectures in Text Summarization

International Conference on Applied Engineering and Natural Sciences

The need for efficient and effective techniques for automatic text summarization has become increasingly critical with the exponential growth of textual data in different domains. Summarizing long texts into short summaries facilitates a quick understanding of the key information contained in the documents. In this paper, we evaluate various architectures for automatic text summarization using the TEDx dataset, a valuable resource consisting of a large collection of TED talks with rich and informative speech transcripts. Our research focuses on evaluating the performance of Long Short-Term Memory (LSTM), Gated Recurrent Unit (GRU), Recurrent Neural Network (RNN) and Transformer architectures for automatic text summarization. We measure the accuracy of each model by comparing the generated summaries with human-written summaries. The findings show that the Transformer model achieves the highest accuracy, followed closely by the GRU model. However, LSTM, RNN exhibit relatively lower ac...

Voice Based Summary Generation using LSTM

International Journal for Research in Applied Science & Engineering Technology (IJRASET), 2022

generation is utilized to have a fast outline of a total meeting without wasting much time. In this research paper, we are making a pipeline to create a summary of a voice recording. We will change the voice recording into text and afterward, we will utilize a deep learning model to produce an outline of that voice recording. We have 10 min voice recording as information and a rundown as result. Rundown age is essential of two sorts: Abstractive synopsis and extractive outline. Abstractive rundown creates new sentences with comparable significance as a synopsis of huge sentences and extractive outline produces sentences in the wake of pulling just significant expressions and catchphrases of enormous sentences.

Summary Generation Using Deep Learning

IEEE EXPLORE, 2021

Summarization is the process of converting any given long text into a lossy compressed version while preserving the overall essence and meaning of the original given text. A problem in natural language processing is the task of summarizing a voluminous text. This research, lays emphasis on abstractive summarization of the text, a process which generates a coherent, short and meaningful text by learning the context from given input source text. This work is a preliminary attempt to generate abstractive summary. The proposed work uses the Attentional Encoder and Decoder based Sequence to Sequence Recurrent Neural Network model. The proposed research's primary objective is to get insights on the working of the mentioned Sequence architecture and the intricacies of different subcomponents of the encoders and the decoders architecture working together so as to give a particular end result. The scope of this paper covers an experimentation of performing text summarization process using Encoder and Decoder based Sequential Learning Recurrent Neural Network design. Satisfactory summary results are obtained after the model is experimented and analyzed on the Amazon Fine Food reviews.

An Optimized Abstractive Text Summarization Model Using Peephole Convolutional LSTM

Symmetry, 2019

ive text summarization that generates a summary by paraphrasing a long text remains an open significant problem for natural language processing. In this paper, we present an abstractive text summarization model, multi-layered attentional peephole convolutional LSTM (long short-term memory) (MAPCoL) that automatically generates a summary from a long text. We optimize parameters of MAPCoL using central composite design (CCD) in combination with the response surface methodology (RSM), which gives the highest accuracy in terms of summary generation. We record the accuracy of our model (MAPCoL) on a CNN/DailyMail dataset. We perform a comparative analysis of the accuracy of MAPCoL with that of the state-of-the-art models in different experimental settings. The MAPCoL also outperforms the traditional LSTM-based models in respect of semantic coherence in the output summary.

MeetingBank: A Benchmark Dataset for Meeting Summarization

arXiv (Cornell University), 2023

As the number of recorded meetings increases, it becomes increasingly important to utilize summarization technology to create useful summaries of these recordings. However, there is a crucial lack of annotated meeting corpora for developing this technology, as it can be hard to collect meetings, especially when the topics discussed are confidential. Furthermore, meeting summaries written by experienced writers are scarce, making it hard for abstractive summarizers to produce sensible output without a reliable reference. This lack of annotated corpora has hindered the development of meeting summarization technology. In this paper, we present MeetingBank, a new benchmark dataset of city council meetings over the past decade. Meeting-Bank is unique among other meeting corpora due to its divide-and-conquer approach, which involves dividing professionally written meeting minutes into shorter passages and aligning them with specific segments of the meeting. This breaks down the process of summarizing a lengthy meeting into smaller, more manageable tasks. The dataset provides a new testbed of various meeting summarization systems and also allows the public to gain insight into how council decisions are made. We make the collection, including meeting video links, transcripts, reference summaries, agenda, and other metadata, publicly available to facilitate the development of better meeting summarization techniques. 1

Abstractive Sentence Summarization with Attentive Recurrent Neural Networks

Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies

ive Sentence Summarization generates a shorter version of a given sentence while attempting to preserve its meaning. We introduce a conditional recurrent neural network (RNN) which generates a summary of an input sentence. The conditioning is provided by a novel convolutional attention-based encoder which ensures that the decoder focuses on the appropriate input words at each step of generation. Our model relies only on learned features and is easy to train in an end-to-end fashion on large data sets. Our experiments show that the model significantly outperforms the recently proposed state-of-the-art method on the Gigaword corpus while performing competitively on the DUC-2004 shared task.

Abstractive Text Summarization using Sequence-to-sequence RNNs and Beyond

Proceedings of The 20th SIGNLL Conference on Computational Natural Language Learning, 2016

In this work, we model abstractive text summarization using Attentional Encoder-Decoder Recurrent Neural Networks, and show that they achieve state-of-the-art performance on two different corpora. We propose several novel models that address critical problems in summarization that are not adequately modeled by the basic architecture, such as modeling keywords , capturing the hierarchy of sentence-toword structure, and emitting words that are rare or unseen at training time. Our work shows that many of our proposed models contribute to further improvement in performance. We also propose a new dataset consisting of multi-sentence summaries, and establish performance benchmarks for further research.

NUTS: Network for Unsupervised Telegraphic Summarization

2018

Extractive summarization methods operate by ranking and selecting the sentences which best encapsulate the theme of a given document. They do not fare well in domains like fictional narratives where there is no central theme and core information is not encapsulated by a small set of sentences. For the purpose of reducing the size of the document while conveying the idea expressed by each sentence, we need more sentence specific methods. Telegraphic summarization, which selects short segments across several sentences, is better suited for such domains. Telegraphic summarization captures the plot better by retaining shorter versions of each sentence while not really concerning itself with grammatically linking these segments. In this paper, we propose an unsupervised deep learning network (NUTS) to generate telegraphic summaries. We use multiple encoderdecoder networks and learn to drop portions of the text that are inferable from the chosen segments. The model is agnostic to both sen...

Deep Learning Based Abstractive Text Summarization: Approaches, Datasets, Evaluation Measures, and Challenges

Mathematical Problems in Engineering

In recent years, the volume of textual data has rapidly increased, which has generated a valuable resource for extracting and analysing information. To retrieve useful knowledge within a reasonable time period, this information must be summarised. This paper reviews recent approaches for abstractive text summarisation using deep learning models. In addition, existing datasets for training and validating these approaches are reviewed, and their features and limitations are presented. The Gigaword dataset is commonly employed for single-sentence summary approaches, while the Cable News Network (CNN)/Daily Mail dataset is commonly employed for multisentence summary approaches. Furthermore, the measures that are utilised to evaluate the quality of summarisation are investigated, and Recall-Oriented Understudy for Gisting Evaluation 1 (ROUGE1), ROUGE2, and ROUGE-L are determined to be the most commonly applied metrics. The challenges that are encountered during the summarisation process ...

Controlling Length in Abstractive Summarization Using a Convolutional Neural Network

Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2018

Convolutional neural networks (CNNs) have met great success in abstractive summarization, but they cannot effectively generate summaries of desired lengths. Because generated summaries are used in difference scenarios which may have space or length constraints, the ability to control the summary length in abstractive summarization is an important problem. In this paper, we propose an approach to constrain the summary length by extending a convolutional sequence to sequence model. The results show that this approach generates high-quality summaries with user defined length, and outperforms the baselines consistently in terms of ROUGE score, length variations and semantic similarity.