Search (18 results, page 1 of 1)

  • × theme_ss:"Automatisches Abstracting"
  1. Goh, A.; Hui, S.C.: TES: a text extraction system (1996) 0.01
    0.014505414 = product of:
      0.04351624 = sum of:
        0.029623196 = weight(_text_:internet in 6599) [ClassicSimilarity], result of:
          0.029623196 = score(doc=6599,freq=2.0), product of:
            0.11352337 = queryWeight, product of:
              2.9522398 = idf(docFreq=6276, maxDocs=44218)
              0.038453303 = queryNorm
            0.2609436 = fieldWeight in 6599, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.9522398 = idf(docFreq=6276, maxDocs=44218)
              0.0625 = fieldNorm(doc=6599)
        0.0138930455 = product of:
          0.041679136 = sum of:
            0.041679136 = weight(_text_:22 in 6599) [ClassicSimilarity], result of:
              0.041679136 = score(doc=6599,freq=2.0), product of:
                0.13465692 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038453303 = queryNorm
                0.30952093 = fieldWeight in 6599, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6599)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    With the onset of the information explosion arising from digital libraries and access to a wealth of information through the Internet, the need to efficiently determine the relevance of a document becomes even more urgent. Describes a text extraction system (TES), which retrieves a set of sentences from a document to form an indicative abstract. Such an automated process enables information to be filtered more quickly. Discusses the combination of various text extraction techniques. Compares results with manually produced abstracts
    Date
    26. 2.1997 10:22:43
  2. Kim, H.H.; Kim, Y.H.: Generic speech summarization of transcribed lecture videos : using tags and their semantic relations (2016) 0.01
    0.005815042 = product of:
      0.034890253 = sum of:
        0.034890253 = product of:
          0.052335378 = sum of:
            0.026285918 = weight(_text_:29 in 2640) [ClassicSimilarity], result of:
              0.026285918 = score(doc=2640,freq=2.0), product of:
                0.13526669 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.038453303 = queryNorm
                0.19432661 = fieldWeight in 2640, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2640)
            0.02604946 = weight(_text_:22 in 2640) [ClassicSimilarity], result of:
              0.02604946 = score(doc=2640,freq=2.0), product of:
                0.13465692 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038453303 = queryNorm
                0.19345059 = fieldWeight in 2640, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2640)
          0.6666667 = coord(2/3)
      0.16666667 = coord(1/6)
    
    Date
    22. 1.2016 12:29:41
  3. Pinto, M.: Engineering the production of meta-information : the abstracting concern (2003) 0.01
    0.005782607 = product of:
      0.03469564 = sum of:
        0.03469564 = product of:
          0.10408692 = sum of:
            0.10408692 = weight(_text_:29 in 4667) [ClassicSimilarity], result of:
              0.10408692 = score(doc=4667,freq=4.0), product of:
                0.13526669 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.038453303 = queryNorm
                0.7694941 = fieldWeight in 4667, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4667)
          0.33333334 = coord(1/3)
      0.16666667 = coord(1/6)
    
    Date
    27.11.2005 18:29:55
    Source
    Journal of information science. 29(2003) no.5, S.405-418
  4. Rodríguez-Vidal, J.; Carrillo-de-Albornoz, J.; Gonzalo, J.; Plaza, L.: Authority and priority signals in automatic summary generation for online reputation management (2021) 0.00
    0.0043639094 = product of:
      0.026183454 = sum of:
        0.026183454 = weight(_text_:internet in 213) [ClassicSimilarity], result of:
          0.026183454 = score(doc=213,freq=4.0), product of:
            0.11352337 = queryWeight, product of:
              2.9522398 = idf(docFreq=6276, maxDocs=44218)
              0.038453303 = queryNorm
            0.23064373 = fieldWeight in 213, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.9522398 = idf(docFreq=6276, maxDocs=44218)
              0.0390625 = fieldNorm(doc=213)
      0.16666667 = coord(1/6)
    
    Abstract
    Online reputation management (ORM) comprises the collection of techniques that help monitoring and improving the public image of an entity (companies, products, institutions) on the Internet. The ORM experts try to minimize the negative impact of the information about an entity while maximizing the positive material for being more trustworthy to the customers. Due to the huge amount of information that is published on the Internet every day, there is a need to summarize the entire flow of information to obtain only those data that are relevant to the entities. Traditionally the automatic summarization task in the ORM scenario takes some in-domain signals into account such as popularity, polarity for reputation and novelty but exists other feature to be considered, the authority of the people. This authority depends on the ability to convince others and therefore to influence opinions. In this work, we propose the use of authority signals that measures the influence of a user jointly with (a) priority signals related to the ORM domain and (b) information regarding the different topics that influential people is talking about. Our results indicate that the use of authority signals may significantly improve the quality of the summaries that are automatically generated.
  5. Endres-Niggemeyer, B.: Summarizing information (1998) 0.00
    0.0037028994 = product of:
      0.022217397 = sum of:
        0.022217397 = weight(_text_:internet in 688) [ClassicSimilarity], result of:
          0.022217397 = score(doc=688,freq=2.0), product of:
            0.11352337 = queryWeight, product of:
              2.9522398 = idf(docFreq=6276, maxDocs=44218)
              0.038453303 = queryNorm
            0.1957077 = fieldWeight in 688, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.9522398 = idf(docFreq=6276, maxDocs=44218)
              0.046875 = fieldNorm(doc=688)
      0.16666667 = coord(1/6)
    
    Abstract
    Summarizing is the process of reducing the large information size of something like a novel or a scientific paper to a short summary or abstract comprising only the most essential points. Summarizing is frequent in everyday communication, but it is also a professional skill for journalists and others. Automated summarizing functions are urgently needed by Internet users who wish to avoid being overwhelmed by information. This book presents the state of the art and surveys related research; it deals with everyday and professional summarizing as well as computerized approaches. The author focuses in detail on the cognitive pro-cess involved in summarizing and supports this with a multimedia simulation systems on the accompanying CD-ROM
  6. Nomoto, T.: Discriminative sentence compression with conditional random fields (2007) 0.00
    0.0037028994 = product of:
      0.022217397 = sum of:
        0.022217397 = weight(_text_:internet in 945) [ClassicSimilarity], result of:
          0.022217397 = score(doc=945,freq=2.0), product of:
            0.11352337 = queryWeight, product of:
              2.9522398 = idf(docFreq=6276, maxDocs=44218)
              0.038453303 = queryNorm
            0.1957077 = fieldWeight in 945, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.9522398 = idf(docFreq=6276, maxDocs=44218)
              0.046875 = fieldNorm(doc=945)
      0.16666667 = coord(1/6)
    
    Abstract
    The paper focuses on a particular approach to automatic sentence compression which makes use of a discriminative sequence classifier known as Conditional Random Fields (CRF). We devise several features for CRF that allow it to incorporate information on nonlinear relations among words. Along with that, we address the issue of data paucity by collecting data from RSS feeds available on the Internet, and turning them into training data for use with CRF, drawing on techniques from biology and information retrieval. We also discuss a recursive application of CRF on the syntactic structure of a sentence as a way of improving the readability of the compression it generates. Experiments found that our approach works reasonably well compared to the state-of-the-art system [Knight, K., & Marcu, D. (2002). Summarization beyond sentence extraction: A probabilistic approach to sentence compression. Artificial Intelligence 139, 91-107.].
  7. Salton, G.; Allan, J.; Buckley, C.; Singhal, A.: Automatic analysis, theme generation, and summarization of machine readable texts (1994) 0.00
    0.0029206579 = product of:
      0.017523946 = sum of:
        0.017523946 = product of:
          0.052571837 = sum of:
            0.052571837 = weight(_text_:29 in 1949) [ClassicSimilarity], result of:
              0.052571837 = score(doc=1949,freq=2.0), product of:
                0.13526669 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.038453303 = queryNorm
                0.38865322 = fieldWeight in 1949, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1949)
          0.33333334 = coord(1/3)
      0.16666667 = coord(1/6)
    
    Date
    16. 8.1998 12:30:29
  8. Craven, T.C.: ¬A phrase flipper for the assistance of writers of abstracts and other text (1995) 0.00
    0.002336526 = product of:
      0.014019156 = sum of:
        0.014019156 = product of:
          0.042057466 = sum of:
            0.042057466 = weight(_text_:29 in 4897) [ClassicSimilarity], result of:
              0.042057466 = score(doc=4897,freq=2.0), product of:
                0.13526669 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.038453303 = queryNorm
                0.31092256 = fieldWeight in 4897, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4897)
          0.33333334 = coord(1/3)
      0.16666667 = coord(1/6)
    
    Date
    17. 8.1996 10:29:59
  9. Robin, J.; McKeown, K.: Empirically designing and evaluating a new revision-based model for summary generation (1996) 0.00
    0.0023155077 = product of:
      0.0138930455 = sum of:
        0.0138930455 = product of:
          0.041679136 = sum of:
            0.041679136 = weight(_text_:22 in 6751) [ClassicSimilarity], result of:
              0.041679136 = score(doc=6751,freq=2.0), product of:
                0.13465692 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038453303 = queryNorm
                0.30952093 = fieldWeight in 6751, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6751)
          0.33333334 = coord(1/3)
      0.16666667 = coord(1/6)
    
    Date
    6. 3.1997 16:22:15
  10. Jones, P.A.; Bradbeer, P.V.G.: Discovery of optimal weights in a concept selection system (1996) 0.00
    0.0023155077 = product of:
      0.0138930455 = sum of:
        0.0138930455 = product of:
          0.041679136 = sum of:
            0.041679136 = weight(_text_:22 in 6974) [ClassicSimilarity], result of:
              0.041679136 = score(doc=6974,freq=2.0), product of:
                0.13465692 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038453303 = queryNorm
                0.30952093 = fieldWeight in 6974, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6974)
          0.33333334 = coord(1/3)
      0.16666667 = coord(1/6)
    
    Source
    Information retrieval: new systems and current research. Proceedings of the 16th Research Colloquium of the British Computer Society Information Retrieval Specialist Group, Drymen, Scotland, 22-23 Mar 94. Ed.: R. Leon
  11. Uyttendaele, C.; Moens, M.-F.; Dumortier, J.: SALOMON: automatic abstracting of legal cases for effective access to court decisions (1998) 0.00
    0.0020444603 = product of:
      0.012266762 = sum of:
        0.012266762 = product of:
          0.036800284 = sum of:
            0.036800284 = weight(_text_:29 in 495) [ClassicSimilarity], result of:
              0.036800284 = score(doc=495,freq=2.0), product of:
                0.13526669 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.038453303 = queryNorm
                0.27205724 = fieldWeight in 495, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=495)
          0.33333334 = coord(1/3)
      0.16666667 = coord(1/6)
    
    Date
    17. 7.1996 14:16:29
  12. Vanderwende, L.; Suzuki, H.; Brockett, J.M.; Nenkova, A.: Beyond SumBasic : task-focused summarization with sentence simplification and lexical expansion (2007) 0.00
    0.0017366307 = product of:
      0.010419784 = sum of:
        0.010419784 = product of:
          0.03125935 = sum of:
            0.03125935 = weight(_text_:22 in 948) [ClassicSimilarity], result of:
              0.03125935 = score(doc=948,freq=2.0), product of:
                0.13465692 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038453303 = queryNorm
                0.23214069 = fieldWeight in 948, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=948)
          0.33333334 = coord(1/3)
      0.16666667 = coord(1/6)
    
    Abstract
    In recent years, there has been increased interest in topic-focused multi-document summarization. In this task, automatic summaries are produced in response to a specific information request, or topic, stated by the user. The system we have designed to accomplish this task comprises four main components: a generic extractive summarization system, a topic-focusing component, sentence simplification, and lexical expansion of topic words. This paper details each of these components, together with experiments designed to quantify their individual contributions. We include an analysis of our results on two large datasets commonly used to evaluate task-focused summarization, the DUC2005 and DUC2006 datasets, using automatic metrics. Additionally, we include an analysis of our results on the DUC2006 task according to human evaluation metrics. In the human evaluation of system summaries compared to human summaries, i.e., the Pyramid method, our system ranked first out of 22 systems in terms of overall mean Pyramid score; and in the human evaluation of summary responsiveness to the topic, our system ranked third out of 35 systems.
  13. Sweeney, S.; Crestani, F.; Losada, D.E.: 'Show me more' : incremental length summarisation using novelty detection (2008) 0.00
    0.0014603289 = product of:
      0.008761973 = sum of:
        0.008761973 = product of:
          0.026285918 = sum of:
            0.026285918 = weight(_text_:29 in 2054) [ClassicSimilarity], result of:
              0.026285918 = score(doc=2054,freq=2.0), product of:
                0.13526669 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.038453303 = queryNorm
                0.19432661 = fieldWeight in 2054, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2054)
          0.33333334 = coord(1/3)
      0.16666667 = coord(1/6)
    
    Date
    29. 7.2008 19:35:12
  14. Wang, S.; Koopman, R.: Embed first, then predict (2019) 0.00
    0.0014603289 = product of:
      0.008761973 = sum of:
        0.008761973 = product of:
          0.026285918 = sum of:
            0.026285918 = weight(_text_:29 in 5400) [ClassicSimilarity], result of:
              0.026285918 = score(doc=5400,freq=2.0), product of:
                0.13526669 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.038453303 = queryNorm
                0.19432661 = fieldWeight in 5400, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5400)
          0.33333334 = coord(1/3)
      0.16666667 = coord(1/6)
    
    Date
    29. 9.2019 12:18:42
  15. Wu, Y.-f.B.; Li, Q.; Bot, R.S.; Chen, X.: Finding nuggets in documents : a machine learning approach (2006) 0.00
    0.0014471923 = product of:
      0.008683153 = sum of:
        0.008683153 = product of:
          0.02604946 = sum of:
            0.02604946 = weight(_text_:22 in 5290) [ClassicSimilarity], result of:
              0.02604946 = score(doc=5290,freq=2.0), product of:
                0.13465692 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038453303 = queryNorm
                0.19345059 = fieldWeight in 5290, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5290)
          0.33333334 = coord(1/3)
      0.16666667 = coord(1/6)
    
    Date
    22. 7.2006 17:25:48
  16. Oh, H.; Nam, S.; Zhu, Y.: Structured abstract summarization of scientific articles : summarization using full-text section information (2023) 0.00
    0.0014471923 = product of:
      0.008683153 = sum of:
        0.008683153 = product of:
          0.02604946 = sum of:
            0.02604946 = weight(_text_:22 in 889) [ClassicSimilarity], result of:
              0.02604946 = score(doc=889,freq=2.0), product of:
                0.13465692 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038453303 = queryNorm
                0.19345059 = fieldWeight in 889, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=889)
          0.33333334 = coord(1/3)
      0.16666667 = coord(1/6)
    
    Date
    22. 1.2023 18:57:12
  17. Jiang, Y.; Meng, R.; Huang, Y.; Lu, W.; Liu, J.: Generating keyphrases for readers : a controllable keyphrase generation framework (2023) 0.00
    0.0014471923 = product of:
      0.008683153 = sum of:
        0.008683153 = product of:
          0.02604946 = sum of:
            0.02604946 = weight(_text_:22 in 1012) [ClassicSimilarity], result of:
              0.02604946 = score(doc=1012,freq=2.0), product of:
                0.13465692 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038453303 = queryNorm
                0.19345059 = fieldWeight in 1012, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1012)
          0.33333334 = coord(1/3)
      0.16666667 = coord(1/6)
    
    Date
    22. 6.2023 14:55:20
  18. Kannan, R.; Ghinea, G.; Swaminathan, S.: What do you wish to see? : A summarization system for movies based on user preferences (2015) 0.00
    0.001168263 = product of:
      0.007009578 = sum of:
        0.007009578 = product of:
          0.021028733 = sum of:
            0.021028733 = weight(_text_:29 in 2683) [ClassicSimilarity], result of:
              0.021028733 = score(doc=2683,freq=2.0), product of:
                0.13526669 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.038453303 = queryNorm
                0.15546128 = fieldWeight in 2683, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2683)
          0.33333334 = coord(1/3)
      0.16666667 = coord(1/6)
    
    Date
    25. 1.2016 18:45:29