Search (33 results, page 1 of 2)

  • × theme_ss:"Automatisches Abstracting"
  • × year_i:[1990 TO 2000}
  1. Ruda, S.: Abstracting: eine Auswahlbibliographie (1992) 0.02
    0.017148167 = product of:
      0.0514445 = sum of:
        0.010820055 = weight(_text_:in in 6603) [ClassicSimilarity], result of:
          0.010820055 = score(doc=6603,freq=6.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.1822149 = fieldWeight in 6603, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0546875 = fieldNorm(doc=6603)
        0.040624447 = weight(_text_:und in 6603) [ClassicSimilarity], result of:
          0.040624447 = score(doc=6603,freq=12.0), product of:
            0.09675359 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.043654136 = queryNorm
            0.41987535 = fieldWeight in 6603, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=6603)
      0.33333334 = coord(2/6)
    
    Abstract
    Die vorliegende Auswahlbibliographie ist in 9 Themenbereiche unterteilt. Der erste Abschnitt enthält Literatur, in der auf Abstracts und Abstracting-Verfahren allgemein eingegangen und ein Überblick über den Stand der Forschung gegeben wird. Im nächsten Abschnitt werden solche Aufsätze referiert, die die historische Entwicklung des Abstracting beschreiben. Im dritten Teil sind Abstracting-Richtlinien verschiedener Institutionen aufgelistet. Lexikalische, syntaktische und semantische Textkondensierungsverfahren sind das Thema der in Abschnitt 4 präsentierten Arbeiten. Textstrukturen von Abstracts werden unter Punkt 5 betrachtet, und die Arbeiten des nächsten Themenbereiches befassen sich mit dem Problem des Schreibens von Abstracts. Der siebte Abschnitt listet sog. 'maschinelle' und maschinen-unterstützte Abstracting-Methoden auf. Anschließend werden 'maschinelle' und maschinenunterstützte Abstracting-Verfahren, Abstracts im Vergleich zu ihren Primärtexten sowie Abstracts im allgemeien bewertet. Den Abschluß bilden Bibliographien
  2. Kuhlen, R.: Abstracts, abstracting : intellektuelle und maschinelle Verfahren (1997) 0.02
    0.016972285 = product of:
      0.050916854 = sum of:
        0.010709076 = weight(_text_:in in 7800) [ClassicSimilarity], result of:
          0.010709076 = score(doc=7800,freq=2.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.18034597 = fieldWeight in 7800, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.09375 = fieldNorm(doc=7800)
        0.040207777 = weight(_text_:und in 7800) [ClassicSimilarity], result of:
          0.040207777 = score(doc=7800,freq=4.0), product of:
            0.09675359 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.043654136 = queryNorm
            0.41556883 = fieldWeight in 7800, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.09375 = fieldNorm(doc=7800)
      0.33333334 = coord(2/6)
    
    Source
    Grundlagen der praktischen Information und Dokumentation: ein Handbuch zur Einführung in die fachliche Informationsarbeit. 4. Aufl. Hrsg.: M. Buder u.a
  3. Endres-Niggemeyer, B.: Bessere Information durch Zusammenfassen aus dem WWW (1999) 0.02
    0.015065096 = product of:
      0.04519529 = sum of:
        0.012365777 = weight(_text_:in in 4496) [ClassicSimilarity], result of:
          0.012365777 = score(doc=4496,freq=6.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.2082456 = fieldWeight in 4496, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=4496)
        0.032829512 = weight(_text_:und in 4496) [ClassicSimilarity], result of:
          0.032829512 = score(doc=4496,freq=6.0), product of:
            0.09675359 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.043654136 = queryNorm
            0.33931053 = fieldWeight in 4496, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=4496)
      0.33333334 = coord(2/6)
    
    Abstract
    Am Beispiel der Knochenmarktransplantation, eines medizinischen Spezialgebietes, wird im folgenden dargelegt, wie man BenutzerInnen eine großen Teil des Aufwandes bei der Wissensbeschaffung abnehmen kann, indem man Suchergebnisse aus dem Netz fragebezogen zusammenfaßt. Dadurch wird in zeitkritischen Situationen, wie sie in Diagnose und Therapie alltäglich sind, die Aufnahme neuen Wissens ermöglicht. Auf einen Überblick über den Stand des Textzusammenfassens und der Ontologieentwicklung folgt eine Systemskizze, in der die Informationssuche im WWW durch ein kognitiv fundiertes Zusammenfassungssystem ergänzt wird. Dazu wird eine Fach-Ontologie vorgeschlagen, die das benötigte Wissen organisiert und repräsentiert.
  4. Endres-Niggemeyer, B.: ¬An empirical process model of abstracting (1992) 0.01
    0.013046755 = product of:
      0.039140265 = sum of:
        0.010709076 = weight(_text_:in in 8834) [ClassicSimilarity], result of:
          0.010709076 = score(doc=8834,freq=2.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.18034597 = fieldWeight in 8834, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.09375 = fieldNorm(doc=8834)
        0.02843119 = weight(_text_:und in 8834) [ClassicSimilarity], result of:
          0.02843119 = score(doc=8834,freq=2.0), product of:
            0.09675359 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.043654136 = queryNorm
            0.29385152 = fieldWeight in 8834, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.09375 = fieldNorm(doc=8834)
      0.33333334 = coord(2/6)
    
    Source
    Mensch und Maschine: Informationelle Schnittstellen der Kommunikation. Proc. des 3. Int. Symposiums für Informationswissenschaft (ISI'92), 5.-7.11.1992 in Saarbrücken. Hrsg.: H.H. Zimmermann, H.-D. Luckhardt u. A. Schulz
  5. Dammeyer, A.; Jürgensen, W.; Krüwel, C.; Poliak, E.; Ruttkowski, S.; Schäfer, Th.; Sirava, M.; Hermes, T.: Videoanalyse mit DiVA (1998) 0.01
    0.013046755 = product of:
      0.039140265 = sum of:
        0.010709076 = weight(_text_:in in 23) [ClassicSimilarity], result of:
          0.010709076 = score(doc=23,freq=8.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.18034597 = fieldWeight in 23, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=23)
        0.02843119 = weight(_text_:und in 23) [ClassicSimilarity], result of:
          0.02843119 = score(doc=23,freq=8.0), product of:
            0.09675359 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.043654136 = queryNorm
            0.29385152 = fieldWeight in 23, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=23)
      0.33333334 = coord(2/6)
    
    Abstract
    Die Bedeutung von Videos nimmt für multimediale Systeme stetig zu. Dabei existiert eine Vielzahl von Produkten zur Betrachtung von Videos, allerdings nur wenige Ansätze, den Inhalt eines Videos zu erschließen. Das DiVA-System, welches an der Universität Bremen im Rahmen eines studentischen Projektes entwickelt wird, dient der automatischen Analyse von MPEG-I Videofilmen. Der dabei verfolgte Ansatz läßt sich in vier Phasen gliedern. Zunächst wird der Videofilm durch eine Shotanalyse in seine einzelnen Kameraeinstellungen (Shots) unterteilt. Darauf aufbauend findet eine Kamerabewegungsanalyse sowie die Erstellung von Mosaicbildern statt. Mit Methoden der künstlichen Intelligenz und der digitalen Bildverarbeitung wird das analysierte Material nach Bild- und Toninformationen ausgewertet. Das Resultat ist eine textuelle Beschreibung eines Videofilms, auf der mit Hilfe von Text-Retrieval-Systemen recherchiert werden kann
    Imprint
    Bremen : Universität Bremen / Fachbereich Mathematik und Informatik
    Source
    Inhaltsbezogene Suche von Bildern und Videosequenzen in digitalen multimedialen Archiven: Beiträge eines Workshops der KI'98 am 16./17.9.1998 in Bremen. Hrsg.: N. Luth
  6. Jones, P.A.; Bradbeer, P.V.G.: Discovery of optimal weights in a concept selection system (1996) 0.01
    0.012645634 = product of:
      0.0379369 = sum of:
        0.014278769 = weight(_text_:in in 6974) [ClassicSimilarity], result of:
          0.014278769 = score(doc=6974,freq=8.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.24046129 = fieldWeight in 6974, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=6974)
        0.02365813 = product of:
          0.04731626 = sum of:
            0.04731626 = weight(_text_:22 in 6974) [ClassicSimilarity], result of:
              0.04731626 = score(doc=6974,freq=2.0), product of:
                0.15286934 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043654136 = queryNorm
                0.30952093 = fieldWeight in 6974, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6974)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Describes the application of weighting strategies to model uncertainties and probabilities in automatic abstracting systems, particularly in the concept selection phase. The weights were originally assigned in an ad hoc manner and were then refined by manual analysis of the results. The new method attempts to derive a more systematic methods and performs this using a genetic algorithm
    Source
    Information retrieval: new systems and current research. Proceedings of the 16th Research Colloquium of the British Computer Society Information Retrieval Specialist Group, Drymen, Scotland, 22-23 Mar 94. Ed.: R. Leon
  7. Robin, J.; McKeown, K.: Empirically designing and evaluating a new revision-based model for summary generation (1996) 0.01
    0.010265838 = product of:
      0.030797515 = sum of:
        0.0071393843 = weight(_text_:in in 6751) [ClassicSimilarity], result of:
          0.0071393843 = score(doc=6751,freq=2.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.120230645 = fieldWeight in 6751, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=6751)
        0.02365813 = product of:
          0.04731626 = sum of:
            0.04731626 = weight(_text_:22 in 6751) [ClassicSimilarity], result of:
              0.04731626 = score(doc=6751,freq=2.0), product of:
                0.15286934 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043654136 = queryNorm
                0.30952093 = fieldWeight in 6751, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6751)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Presents a system for summarizing quantitative data in natural language, focusing on the use of a corpus of basketball game summaries, drawn from online news services, to empirically shape the system design and to evaluate the approach. Initial corpus analysis revealed characteristics of textual summaries that challenge the capabilities of current language generation systems. A revision based corpus analysis was used to identify and encode the revision rules of the system. Presents a quantitative evaluation, using several test corpora, to measure the robustness of the new revision based model
    Date
    6. 3.1997 16:22:15
  8. Kuhlen, R.: Abstracts, abstracting : intellektuelle und maschinelle Verfahren (1990) 0.01
    0.007818179 = product of:
      0.04690907 = sum of:
        0.04690907 = weight(_text_:und in 2333) [ClassicSimilarity], result of:
          0.04690907 = score(doc=2333,freq=4.0), product of:
            0.09675359 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.043654136 = queryNorm
            0.4848303 = fieldWeight in 2333, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.109375 = fieldNorm(doc=2333)
      0.16666667 = coord(1/6)
    
    Source
    Grundlagen der praktischen Information und Dokumentation. 3. Aufl. Hrsg.: M. Buder u.a. Bd.1
  9. Ruda, S.: Maschinenunterstützte Kondensierung von Fachtexten mit CONNY : Abstracting am Beispiel eines 'Nachrichten für Dokumentation'-Textkorpus (1994) 0.00
    0.004787637 = product of:
      0.028725822 = sum of:
        0.028725822 = weight(_text_:und in 2376) [ClassicSimilarity], result of:
          0.028725822 = score(doc=2376,freq=6.0), product of:
            0.09675359 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.043654136 = queryNorm
            0.2968967 = fieldWeight in 2376, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2376)
      0.16666667 = coord(1/6)
    
    Abstract
    Als Textkorpus sind von 50 verschiedenen Autoren verfaßte Dokumente der Zeitschrift 'Nachrichten für Dokumentation' aus einem Zwanzigjahreszeitraum (1969-1989) herangezogen worden. Die Untersuchung der Abstracts hat ergeben, daß lediglich 15 von 50 Abstracts aus ausschließlich 'normgerechten' Abstractsätzen bestehen und kein Abstract allen Anforderungen der Richtlinien genügt. Insofern signalisieren sie die Abstracting-Richtlinien als 'Wunschdenken', was die Idee des maschinenunterstützten Abstracting nach linguistischen Merkmalen bekräftigt. CONNY ist ein interaktives linhuistisches Abstracting-Modell für Fachtexte, das dem Abstractor auf der Oberflächenstruktur operierende allgemeine Abstracting-Richtlinien anbietet. Es kondendiert die als abstractrelevant bewertenden Primärtextteile auf Primärtext-, Satz- und Abstractebene hinsichtlich Lexik, Syntax und Semantik
  10. Goh, A.; Hui, S.C.: TES: a text extraction system (1996) 0.00
    0.003943022 = product of:
      0.02365813 = sum of:
        0.02365813 = product of:
          0.04731626 = sum of:
            0.04731626 = weight(_text_:22 in 6599) [ClassicSimilarity], result of:
              0.04731626 = score(doc=6599,freq=2.0), product of:
                0.15286934 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043654136 = queryNorm
                0.30952093 = fieldWeight in 6599, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6599)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Date
    26. 2.1997 10:22:43
  11. Johnson, F.: Automatic abstracting research (1995) 0.00
    0.0023797948 = product of:
      0.014278769 = sum of:
        0.014278769 = weight(_text_:in in 3847) [ClassicSimilarity], result of:
          0.014278769 = score(doc=3847,freq=8.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.24046129 = fieldWeight in 3847, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=3847)
      0.16666667 = coord(1/6)
    
    Abstract
    Discusses the attraction for researchers of the prospect of automatically generating abstracts but notes that the promise of superseding the human effort has yet to be realized. Notes ways in which progress in automatic abstracting research may come about and suggests a shift in the aim from reproducing the conventional benefits of abstracts to accentuating the advantages to users of the computerized representation of information in large textual databases
  12. Johnson, F.C.; Paice, C.D.; Black, W.J.; Neal, A.P.: ¬The application of linguistic processing to automatic abstract generation (1993) 0.00
    0.0021034614 = product of:
      0.012620768 = sum of:
        0.012620768 = weight(_text_:in in 2290) [ClassicSimilarity], result of:
          0.012620768 = score(doc=2290,freq=4.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.21253976 = fieldWeight in 2290, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.078125 = fieldNorm(doc=2290)
      0.16666667 = coord(1/6)
    
    Footnote
    Wiederabgedruckt in: Readings in information retrieval. Ed.: K. Sparck Jones u. P. Willett. San Francisco: Morgan Kaufmann 1997. S.538-552.
  13. Salton, G.; Allan, J.; Buckley, C.; Singhal, A.: Automatic analysis, theme generation, and summarization of machine readable texts (1994) 0.00
    0.0021034614 = product of:
      0.012620768 = sum of:
        0.012620768 = weight(_text_:in in 1949) [ClassicSimilarity], result of:
          0.012620768 = score(doc=1949,freq=4.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.21253976 = fieldWeight in 1949, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.078125 = fieldNorm(doc=1949)
      0.16666667 = coord(1/6)
    
    Footnote
    Wiederabgedruckt in: Readings in information retrieval. Ed.: K. Sparck Jones u. P. Willett. San Francisco: Morgan Kaufmann 1997. S.478-483.
  14. Advances in automatic text summarization (1999) 0.00
    0.0021034614 = product of:
      0.012620768 = sum of:
        0.012620768 = weight(_text_:in in 6191) [ClassicSimilarity], result of:
          0.012620768 = score(doc=6191,freq=4.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.21253976 = fieldWeight in 6191, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.078125 = fieldNorm(doc=6191)
      0.16666667 = coord(1/6)
    
    Footnote
    Rez. in: Knowledge organization 27(2000) no.3, S.178-180 (H. Saggion)
  15. Bateman, J.; Teich, E.: Selective information presentation in an integrated publication system : an application of genre-driven text generation (1995) 0.00
    0.0020823204 = product of:
      0.012493922 = sum of:
        0.012493922 = weight(_text_:in in 2928) [ClassicSimilarity], result of:
          0.012493922 = score(doc=2928,freq=2.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.21040362 = fieldWeight in 2928, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.109375 = fieldNorm(doc=2928)
      0.16666667 = coord(1/6)
    
  16. Ahmad, K.: Text summarisation : the role of lexical cohesion analysis (1995) 0.00
    0.0020609628 = product of:
      0.012365777 = sum of:
        0.012365777 = weight(_text_:in in 5795) [ClassicSimilarity], result of:
          0.012365777 = score(doc=5795,freq=6.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.2082456 = fieldWeight in 5795, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=5795)
      0.16666667 = coord(1/6)
    
    Abstract
    The work in automatic text summary focuses mainly on computational models of texts. The artificial intelligence related work in text summary deals mainly with narrative texts such as newspaper reports and stories. Presents a study on the summary of non-narrative texts such as those in scientific and technical communication. Discusses syntactic cohesion; lexical cohesion; complex lexical repetition; simple and complex paraphrase; bonds and links; and Tele-pattan; an architecture for cohesion based text analysis and summarisation system working on SGML
  17. Moens, M.-F.; Uyttendaele, C.; Dumotier, J.: Abstracting of legal cases : the potential of clustering based on the selection of representative objects (1999) 0.00
    0.0019955188 = product of:
      0.011973113 = sum of:
        0.011973113 = weight(_text_:in in 2944) [ClassicSimilarity], result of:
          0.011973113 = score(doc=2944,freq=10.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.20163295 = fieldWeight in 2944, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=2944)
      0.16666667 = coord(1/6)
    
    Abstract
    The SALOMON project automatically summarizes Belgian criminal cases in order to improve access to the large number of existing and future court decisions. SALOMON extracts text units from the case text to form a case summary. Such a case summary facilitates the rapid determination of the relevance of the case or may be employed in text search. an important part of the research concerns the development of techniques for automatic recognition of representative text paragraphs (or sentences) in texts of unrestricted domains. these techniques are employed to eliminate redundant material in the case texts, and to identify informative text paragraphs which are relevant to include in the case summary. An evaluation of a test set of 700 criminal cases demonstrates that the algorithms have an application potential for automatic indexing, abstracting, and text linkage
  18. Moens, M.-F.; Uyttendaele, C.: Automatic text structuring and categorization as a first step in summarizing legal cases (1997) 0.00
    0.0017848461 = product of:
      0.010709076 = sum of:
        0.010709076 = weight(_text_:in in 2256) [ClassicSimilarity], result of:
          0.010709076 = score(doc=2256,freq=8.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.18034597 = fieldWeight in 2256, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=2256)
      0.16666667 = coord(1/6)
    
    Abstract
    The SALOMON system automatically summarizes Belgian criminal cases in order to improve access to the large number of existing and future court decisions. SALOMON extracts relevant text units from the case text to form a case summary. Such a case profile facilitates the rapid determination of the relevance of the case or may be employed in text search. In a first important abstracting step SALOMON performs an initial categorization of legal criminal cases and structures the case text into separate legally relevant and irrelevant components. A text grammar represented as a semantic network is used to automatically determine the category of the case and its components. Extracts from the case general data and identifies text portions relevant for further abstracting. Prior knowledge of the text structure and its indicative cues may support automatic abstracting. A text grammar is a promising form for representing the knowledge involved
  19. Craven, T.C.: ¬A computer-aided abstracting tool kit (1993) 0.00
    0.001682769 = product of:
      0.010096614 = sum of:
        0.010096614 = weight(_text_:in in 6506) [ClassicSimilarity], result of:
          0.010096614 = score(doc=6506,freq=4.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.17003182 = fieldWeight in 6506, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=6506)
      0.16666667 = coord(1/6)
    
    Abstract
    Describes the abstracting assistance features being prototyped in the TEXNET text network management system. Sentence weighting methods include: weithing negatively or positively on the stems in a selected passage; weighting on general lists of cue words, adjusting weights of selected segments; and weighting of occurrence of frequent stems. The user may adjust a number of parameters: the minimum strength of extracts; the threshold for frequent word/stems and the amount sentence weight is to be adjusted for each weighting type
  20. Craven, T.C.: ¬A phrase flipper for the assistance of writers of abstracts and other text (1995) 0.00
    0.001682769 = product of:
      0.010096614 = sum of:
        0.010096614 = weight(_text_:in in 4897) [ClassicSimilarity], result of:
          0.010096614 = score(doc=4897,freq=4.0), product of:
            0.059380736 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.043654136 = queryNorm
            0.17003182 = fieldWeight in 4897, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=4897)
      0.16666667 = coord(1/6)
    
    Abstract
    Describes computerized tools for computer assisted abstracting. FlipPhr is a Microsoft Windows application program that rearranges (flips) phrases or other expressions in accordance with rules in a grammar. The flipping may be invoked with a single keystroke from within various Windows application programs that allow cutting and pasting of text. The user may modify the grammar to provide for different kinds of flipping