Search (31 results, page 1 of 2)

  • × theme_ss:"Automatisches Abstracting"
  • × year_i:[1990 TO 2000}
  1. Summarising software for publishing (1996) 0.06
    0.06390515 = product of:
      0.19171545 = sum of:
        0.014666359 = weight(_text_:of in 5121) [ClassicSimilarity], result of:
          0.014666359 = score(doc=5121,freq=6.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.23940048 = fieldWeight in 5121, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=5121)
        0.09439332 = weight(_text_:software in 5121) [ClassicSimilarity], result of:
          0.09439332 = score(doc=5121,freq=6.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.6073436 = fieldWeight in 5121, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0625 = fieldNorm(doc=5121)
        0.082655765 = product of:
          0.16531153 = sum of:
            0.16531153 = weight(_text_:packages in 5121) [ClassicSimilarity], result of:
              0.16531153 = score(doc=5121,freq=2.0), product of:
                0.2706874 = queryWeight, product of:
                  6.9093957 = idf(docFreq=119, maxDocs=44218)
                  0.03917671 = queryNorm
                0.6107101 = fieldWeight in 5121, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  6.9093957 = idf(docFreq=119, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5121)
          0.5 = coord(1/2)
      0.33333334 = coord(3/9)
    
    Abstract
    Reviews 4 software packages designed to provide accurate and indicative summaries of documents by taking the documents and creating distinctive abstracts from them. The products reviewed are: Oracle's ConText; InText's Object Analyzer; Iconovex's AnchorPage; and Software Scientific's Interrogator. Techniques used by the products include: the use of dictionaries of known words and phrases to interpret documents; and heuristic analysis involving weighting all the words in the document solely on their occurrence and position within the document
  2. McKeown, K.; Robin, J.; Kukich, K.: Generating concise natural language summaries (1995) 0.05
    0.045119576 = product of:
      0.13535872 = sum of:
        0.08389453 = weight(_text_:applications in 2932) [ClassicSimilarity], result of:
          0.08389453 = score(doc=2932,freq=2.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.4864132 = fieldWeight in 2932, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.078125 = fieldNorm(doc=2932)
        0.010584532 = weight(_text_:of in 2932) [ClassicSimilarity], result of:
          0.010584532 = score(doc=2932,freq=2.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.17277241 = fieldWeight in 2932, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.078125 = fieldNorm(doc=2932)
        0.040879667 = weight(_text_:systems in 2932) [ClassicSimilarity], result of:
          0.040879667 = score(doc=2932,freq=2.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.339541 = fieldWeight in 2932, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.078125 = fieldNorm(doc=2932)
      0.33333334 = coord(3/9)
    
    Abstract
    Description of the problems for summary generation, the applications developed (for basket ball games - STREAK and for telephone network planning activity - PLANDOC), the linguistic constructions that the systems use to convey information concisely and the textual constraints that determine what information gets included
  3. Jones, P.A.; Bradbeer, P.V.G.: Discovery of optimal weights in a concept selection system (1996) 0.03
    0.02880529 = product of:
      0.086415865 = sum of:
        0.018934188 = weight(_text_:of in 6974) [ClassicSimilarity], result of:
          0.018934188 = score(doc=6974,freq=10.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.3090647 = fieldWeight in 6974, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=6974)
        0.046250064 = weight(_text_:systems in 6974) [ClassicSimilarity], result of:
          0.046250064 = score(doc=6974,freq=4.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.38414678 = fieldWeight in 6974, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.0625 = fieldNorm(doc=6974)
        0.021231614 = product of:
          0.042463228 = sum of:
            0.042463228 = weight(_text_:22 in 6974) [ClassicSimilarity], result of:
              0.042463228 = score(doc=6974,freq=2.0), product of:
                0.13719016 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03917671 = queryNorm
                0.30952093 = fieldWeight in 6974, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6974)
          0.5 = coord(1/2)
      0.33333334 = coord(3/9)
    
    Abstract
    Describes the application of weighting strategies to model uncertainties and probabilities in automatic abstracting systems, particularly in the concept selection phase. The weights were originally assigned in an ad hoc manner and were then refined by manual analysis of the results. The new method attempts to derive a more systematic methods and performs this using a genetic algorithm
    Source
    Information retrieval: new systems and current research. Proceedings of the 16th Research Colloquium of the British Computer Society Information Retrieval Specialist Group, Drymen, Scotland, 22-23 Mar 94. Ed.: R. Leon
  4. Robin, J.; McKeown, K.: Empirically designing and evaluating a new revision-based model for summary generation (1996) 0.02
    0.024892237 = product of:
      0.07467671 = sum of:
        0.020741362 = weight(_text_:of in 6751) [ClassicSimilarity], result of:
          0.020741362 = score(doc=6751,freq=12.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.33856338 = fieldWeight in 6751, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=6751)
        0.03270373 = weight(_text_:systems in 6751) [ClassicSimilarity], result of:
          0.03270373 = score(doc=6751,freq=2.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.2716328 = fieldWeight in 6751, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.0625 = fieldNorm(doc=6751)
        0.021231614 = product of:
          0.042463228 = sum of:
            0.042463228 = weight(_text_:22 in 6751) [ClassicSimilarity], result of:
              0.042463228 = score(doc=6751,freq=2.0), product of:
                0.13719016 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03917671 = queryNorm
                0.30952093 = fieldWeight in 6751, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6751)
          0.5 = coord(1/2)
      0.33333334 = coord(3/9)
    
    Abstract
    Presents a system for summarizing quantitative data in natural language, focusing on the use of a corpus of basketball game summaries, drawn from online news services, to empirically shape the system design and to evaluate the approach. Initial corpus analysis revealed characteristics of textual summaries that challenge the capabilities of current language generation systems. A revision based corpus analysis was used to identify and encode the revision rules of the system. Presents a quantitative evaluation, using several test corpora, to measure the robustness of the new revision based model
    Date
    6. 3.1997 16:22:15
  5. Craven, T.C.: ¬An experiment in the use of tools for computer-assisted abstracting (1996) 0.02
    0.016579175 = product of:
      0.07460629 = sum of:
        0.016802425 = weight(_text_:of in 7426) [ClassicSimilarity], result of:
          0.016802425 = score(doc=7426,freq=14.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.2742677 = fieldWeight in 7426, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=7426)
        0.05780387 = weight(_text_:software in 7426) [ClassicSimilarity], result of:
          0.05780387 = score(doc=7426,freq=4.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.3719205 = fieldWeight in 7426, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=7426)
      0.22222222 = coord(2/9)
    
    Abstract
    Experimental subjects wrote abstracts of an article using a simplified version of the TEXNET abstracting assistance software. In addition to the fulltext, the 35 subjects were presented with either keywords or phrases extracted automatically. The resulting abstracts, and the times taken, were recorded automatically; some additional information was gathered by oral questionnaire. Results showed considerable variation among subjects, but 37% found the keywords or phrases quite or very useful in writing their abstracts. Statistical analysis failed to support deveral hypothesised relations; phrases were not viewed as significantly more helpful than keywords; and abstracting experience did not correlate with originality of wording, approximation of the author abstract, or greater conciseness. Results also suggested possible modifications to the software
    Source
    Global complexity: information, chaos and control. Proceedings of the 59th Annual Meeting of the American Society for Information Science, ASIS'96, Baltimore, Maryland, 21-24 Oct 1996. Ed.: S. Hardin
  6. Brandow, R.; Mitze, K.; Rau, L.F.: Automatic condensation of electronic publications by sentence selection (1995) 0.01
    0.012245992 = product of:
      0.055106964 = sum of:
        0.022403233 = weight(_text_:of in 2929) [ClassicSimilarity], result of:
          0.022403233 = score(doc=2929,freq=14.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.36569026 = fieldWeight in 2929, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=2929)
        0.03270373 = weight(_text_:systems in 2929) [ClassicSimilarity], result of:
          0.03270373 = score(doc=2929,freq=2.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.2716328 = fieldWeight in 2929, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.0625 = fieldNorm(doc=2929)
      0.22222222 = coord(2/9)
    
    Abstract
    Description of a system that performs domain-independent automatic condensation of news from a large commercial news service encompassing 41 different publications. This system was evaluated against a system that condensed the same articles using only the first portions of the texts (the löead), up to the target length of the summaries. 3 lengths of articles were evaluated for 250 documents by both systems, totalling 1.500 suitability judgements in all. The lead-based summaries outperformed the 'intelligent' summaries significantly, achieving acceptability ratings of over 90%, compared to 74,7%
  7. Goh, A.; Hui, S.C.: TES: a text extraction system (1996) 0.01
    0.008925734 = product of:
      0.040165804 = sum of:
        0.018934188 = weight(_text_:of in 6599) [ClassicSimilarity], result of:
          0.018934188 = score(doc=6599,freq=10.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.3090647 = fieldWeight in 6599, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=6599)
        0.021231614 = product of:
          0.042463228 = sum of:
            0.042463228 = weight(_text_:22 in 6599) [ClassicSimilarity], result of:
              0.042463228 = score(doc=6599,freq=2.0), product of:
                0.13719016 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03917671 = queryNorm
                0.30952093 = fieldWeight in 6599, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6599)
          0.5 = coord(1/2)
      0.22222222 = coord(2/9)
    
    Abstract
    With the onset of the information explosion arising from digital libraries and access to a wealth of information through the Internet, the need to efficiently determine the relevance of a document becomes even more urgent. Describes a text extraction system (TES), which retrieves a set of sentences from a document to form an indicative abstract. Such an automated process enables information to be filtered more quickly. Discusses the combination of various text extraction techniques. Compares results with manually produced abstracts
    Date
    26. 2.1997 10:22:43
  8. Endres-Niggemeyer, B.: Summarizing information (1998) 0.01
    0.007895015 = product of:
      0.03552757 = sum of:
        0.010999769 = weight(_text_:of in 688) [ClassicSimilarity], result of:
          0.010999769 = score(doc=688,freq=6.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.17955035 = fieldWeight in 688, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=688)
        0.0245278 = weight(_text_:systems in 688) [ClassicSimilarity], result of:
          0.0245278 = score(doc=688,freq=2.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.2037246 = fieldWeight in 688, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.046875 = fieldNorm(doc=688)
      0.22222222 = coord(2/9)
    
    Abstract
    Summarizing is the process of reducing the large information size of something like a novel or a scientific paper to a short summary or abstract comprising only the most essential points. Summarizing is frequent in everyday communication, but it is also a professional skill for journalists and others. Automated summarizing functions are urgently needed by Internet users who wish to avoid being overwhelmed by information. This book presents the state of the art and surveys related research; it deals with everyday and professional summarizing as well as computerized approaches. The author focuses in detail on the cognitive pro-cess involved in summarizing and supports this with a multimedia simulation systems on the accompanying CD-ROM
  9. Su, H.: Automatic abstracting (1996) 0.00
    0.0031115601 = product of:
      0.028004041 = sum of:
        0.028004041 = weight(_text_:of in 150) [ClassicSimilarity], result of:
          0.028004041 = score(doc=150,freq=14.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.45711282 = fieldWeight in 150, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.078125 = fieldNorm(doc=150)
      0.11111111 = coord(1/9)
    
    Abstract
    Presents an introductory overview of research into the automatic construction of abstracts from the texts of documents. Discusses the origin and definition of automatic abstracting; reasons for using automatic abstracting; methods of automatic abstracting; and evaluation problems
    Source
    Bulletin of the Library Association of China. 1996, no.56, Jun., S.41-47
  10. Endres-Niggemeyer, B.; Neugebauer, E.: Professional summarizing : no cognitive simulation without observation (1998) 0.00
    0.002661118 = product of:
      0.023950063 = sum of:
        0.023950063 = weight(_text_:of in 3243) [ClassicSimilarity], result of:
          0.023950063 = score(doc=3243,freq=16.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.39093933 = fieldWeight in 3243, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=3243)
      0.11111111 = coord(1/9)
    
    Abstract
    Develops a cognitive model of expert summarization, using 54 working processes of 6 experts recorded by thinking-alound protocols. It comprises up to 140 working steps. Components of the model are a toolbox of empirically founded strategies, principles of process organization, and interpreted working steps where the interaction of cognitive strategies can be investigated. In the computerized simulation the SimSum (Simulation of Summarizing) system, cognitive strategies are represented by object-oriented agents grouped around dedicated blckboards
    Source
    Journal of the American Society for Information Science. 49(1998) no.6, S.486-506
  11. Xianghao, G.; Yixin, Z.; Li, Y.: ¬A new method of news test understanding and abstracting based on speech acts theory (1998) 0.00
    0.002661118 = product of:
      0.023950063 = sum of:
        0.023950063 = weight(_text_:of in 3532) [ClassicSimilarity], result of:
          0.023950063 = score(doc=3532,freq=16.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.39093933 = fieldWeight in 3532, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=3532)
      0.11111111 = coord(1/9)
    
    Abstract
    Presents a method for the automated analysis and comprehension of foreign affairs news produced by a Chinese news agency. Notes that the development of the method was prededed by a study of the structuring rules of the news. Describes how an abstract of the news story is produced automatically from the analysis. Stresses the main aim of the work which is to use specch act theory to analyse and classify sentences
    Source
    Journal of the China Society for Scientific and Technical Information. 17(1998) no.4, S.257-262
  12. Moens, M.-F.; Uyttendaele, C.; Dumotier, J.: Abstracting of legal cases : the potential of clustering based on the selection of representative objects (1999) 0.00
    0.002544205 = product of:
      0.022897845 = sum of:
        0.022897845 = weight(_text_:of in 2944) [ClassicSimilarity], result of:
          0.022897845 = score(doc=2944,freq=26.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.37376386 = fieldWeight in 2944, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=2944)
      0.11111111 = coord(1/9)
    
    Abstract
    The SALOMON project automatically summarizes Belgian criminal cases in order to improve access to the large number of existing and future court decisions. SALOMON extracts text units from the case text to form a case summary. Such a case summary facilitates the rapid determination of the relevance of the case or may be employed in text search. an important part of the research concerns the development of techniques for automatic recognition of representative text paragraphs (or sentences) in texts of unrestricted domains. these techniques are employed to eliminate redundant material in the case texts, and to identify informative text paragraphs which are relevant to include in the case summary. An evaluation of a test set of 700 criminal cases demonstrates that the algorithms have an application potential for automatic indexing, abstracting, and text linkage
    Source
    Journal of the American Society for Information Science. 50(1999) no.2, S.151-161
  13. Craven, T.C.: ¬A computer-aided abstracting tool kit (1993) 0.00
    0.002489248 = product of:
      0.022403233 = sum of:
        0.022403233 = weight(_text_:of in 6506) [ClassicSimilarity], result of:
          0.022403233 = score(doc=6506,freq=14.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.36569026 = fieldWeight in 6506, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=6506)
      0.11111111 = coord(1/9)
    
    Abstract
    Describes the abstracting assistance features being prototyped in the TEXNET text network management system. Sentence weighting methods include: weithing negatively or positively on the stems in a selected passage; weighting on general lists of cue words, adjusting weights of selected segments; and weighting of occurrence of frequent stems. The user may adjust a number of parameters: the minimum strength of extracts; the threshold for frequent word/stems and the amount sentence weight is to be adjusted for each weighting type
    Source
    Canadian journal of information and library science. 18(1993) no.2, S.20-31
  14. Maybury, M.T.: Generating summaries from event data (1995) 0.00
    0.0024443932 = product of:
      0.021999538 = sum of:
        0.021999538 = weight(_text_:of in 2349) [ClassicSimilarity], result of:
          0.021999538 = score(doc=2349,freq=24.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.3591007 = fieldWeight in 2349, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=2349)
      0.11111111 = coord(1/9)
    
    Abstract
    Summarization entails analysis of source material, selection of key information, condensation of this, and generation of a compct summary form. While there habe been many investigations into the automatic summarization of text, relatively little attention has been given to the summarization of information from structured information sources such as data of knowledge bases, despite this being a desirable capability for a number of application areas including report generation from databases (e.g. weather, financial, medical) and simulation (e.g. military, manufacturing, aconomic). After a brief introduction indicating the main elements of summarization and referring to some illustrative approaches to it, considers pecific issues in the generation of text summaries of event data, describes a system, SumGen, which selects key information from an event database by reasoning about event frequencies, frequencies of relations between events, and domain specific importance measures. Describes how Sum Gen then aggregates similar information and plans a summary presentations tailored to stereotypical users
  15. Johnson, F.C.: ¬A critical view of system-centered to user-centered evaluation of automatic abstracting research (1999) 0.00
    0.0024443932 = product of:
      0.021999538 = sum of:
        0.021999538 = weight(_text_:of in 2994) [ClassicSimilarity], result of:
          0.021999538 = score(doc=2994,freq=6.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.3591007 = fieldWeight in 2994, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.09375 = fieldNorm(doc=2994)
      0.11111111 = coord(1/9)
    
    Source
    New review of information and library research. 5(1999), S.49-63
  16. Paice, C.D.: Automatic abstracting (1994) 0.00
    0.0023521183 = product of:
      0.021169065 = sum of:
        0.021169065 = weight(_text_:of in 917) [ClassicSimilarity], result of:
          0.021169065 = score(doc=917,freq=8.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.34554482 = fieldWeight in 917, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.078125 = fieldNorm(doc=917)
      0.11111111 = coord(1/9)
    
    Abstract
    The final report of the 2nd British Library abstracting project (the BLAB project), 1990-1992, which was carried out partly at the Computing Department of Lancaster University, and partly at the Centre for Computational Linguistics, UMIST. This project built on the results of the first project, of 1985-1987, to build a system designed create abstracts automatically from given texts
  17. Uyttendaele, C.; Moens, M.-F.; Dumortier, J.: SALOMON: automatic abstracting of legal cases for effective access to court decisions (1998) 0.00
    0.0023284785 = product of:
      0.020956306 = sum of:
        0.020956306 = weight(_text_:of in 495) [ClassicSimilarity], result of:
          0.020956306 = score(doc=495,freq=16.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.34207192 = fieldWeight in 495, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=495)
      0.11111111 = coord(1/9)
    
    Abstract
    The SALOMON project summarises Belgian criminal cases in order to improve access to the large number of existing and future cases. A double methodology was used when developing SALOMON: the cases are processed by employing additional knowledge to interpret structural patterns and features on the one hand and by way of occurrence statistics of index terms on the other. SALOMON performs an initial categorisation and structuring of the cases and subsequently extracts the most relevant text units of the alleged offences and of the opinion of the court. The SALOMON techniques do not themselves solve any legal questions, but they do guide the use effectively towards relevant texts
  18. Johnson, F.: Automatic abstracting research (1995) 0.00
    0.002304596 = product of:
      0.020741362 = sum of:
        0.020741362 = weight(_text_:of in 3847) [ClassicSimilarity], result of:
          0.020741362 = score(doc=3847,freq=12.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.33856338 = fieldWeight in 3847, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=3847)
      0.11111111 = coord(1/9)
    
    Abstract
    Discusses the attraction for researchers of the prospect of automatically generating abstracts but notes that the promise of superseding the human effort has yet to be realized. Notes ways in which progress in automatic abstracting research may come about and suggests a shift in the aim from reproducing the conventional benefits of abstracts to accentuating the advantages to users of the computerized representation of information in large textual databases
  19. Endres-Niggemeyer, B.; Maier, E.; Sigel, A.: How to implement a naturalistic model of abstracting : four core working steps of an expert abstractor (1995) 0.00
    0.0021780923 = product of:
      0.01960283 = sum of:
        0.01960283 = weight(_text_:of in 2930) [ClassicSimilarity], result of:
          0.01960283 = score(doc=2930,freq=14.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.31997898 = fieldWeight in 2930, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2930)
      0.11111111 = coord(1/9)
    
    Abstract
    4 working steps taken from a comprehensive empirical model of expert abstracting are studied in order to prepare an explorative implementation of a simulation model. It aims at explaining the knowledge processing activities during professional summarizing. Following the case-based and holistic strategy of qualitative empirical research, the main features of the simulation system were developed by investigating in detail a small but central test case - 4 working steps where an expert abstractor discovers what the paper is about and drafts the topic sentence of the abstract
  20. Salton, G.: Automatic text structuring and summarization (1997) 0.00
    0.0021780923 = product of:
      0.01960283 = sum of:
        0.01960283 = weight(_text_:of in 145) [ClassicSimilarity], result of:
          0.01960283 = score(doc=145,freq=14.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.31997898 = fieldWeight in 145, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=145)
      0.11111111 = coord(1/9)
    
    Abstract
    Applies the ideas from the automatic link generation research to automatic text summarisation. Using techniques for inter-document link generation, generates intra-document links between passages of a document. Based on the intra-document linkage pattern of a text, characterises the structure of the text. Applies the knowledge of text structure to do automatic text summarisation by passage extraction. Evaluates a set of 50 summaries generated using these techniques by comparing the to paragraph extracts constructed by humans. The automatic summarisation methods perform well, especially in view of the fact that the summaries generates by 2 humans for the same article are surprisingly dissimilar
    Footnote
    Contribution to a special issue on methods and tools for the automatic construction of hypertext