Search (11 results, page 1 of 1)

  • × language_ss:"e"
  • × theme_ss:"Automatisches Indexieren"
  • × theme_ss:"Computerlinguistik"
  1. Driscoll, J.R.; Rajala, D.A.; Shaffer, W.H.: ¬The operation and performance of an artificially intelligent keywording system (1991) 0.02
    0.023606706 = product of:
      0.18885365 = sum of:
        0.062951215 = product of:
          0.12590243 = sum of:
            0.12590243 = weight(_text_:rules in 6681) [ClassicSimilarity], result of:
              0.12590243 = score(doc=6681,freq=8.0), product of:
                0.16161752 = queryWeight, product of:
                  5.036312 = idf(docFreq=780, maxDocs=44218)
                  0.032090448 = queryNorm
                0.77901477 = fieldWeight in 6681, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  5.036312 = idf(docFreq=780, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=6681)
          0.5 = coord(1/2)
        0.12590243 = weight(_text_:rules in 6681) [ClassicSimilarity], result of:
          0.12590243 = score(doc=6681,freq=8.0), product of:
            0.16161752 = queryWeight, product of:
              5.036312 = idf(docFreq=780, maxDocs=44218)
              0.032090448 = queryNorm
            0.77901477 = fieldWeight in 6681, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              5.036312 = idf(docFreq=780, maxDocs=44218)
              0.0546875 = fieldNorm(doc=6681)
      0.125 = coord(2/16)
    
    Abstract
    Presents a new approach to text analysis for automating the key phrase indexing process, using artificial intelligence techniques. This mimics the behaviour of human experts by using a rule base consisting of insertion and deletion rules generated by subject-matter experts. The insertion rules are based on the idea that some phrases found in a text imply or trigger other phrases. The deletion rules apply to semantically ambiguous phrases where text presence alone does not determine appropriateness as a key phrase. The insertion and deletion rules are used to transform a list of found phrases to a list of key phrases for indexing a document. Statistical data are provided to demonstrate the performance of this expert rule based system
  2. Salton, G.: Automatic processing of foreign language documents (1985) 0.00
    0.0031814466 = product of:
      0.025451573 = sum of:
        0.016484896 = weight(_text_:american in 3650) [ClassicSimilarity], result of:
          0.016484896 = score(doc=3650,freq=2.0), product of:
            0.10940785 = queryWeight, product of:
              3.4093587 = idf(docFreq=3973, maxDocs=44218)
              0.032090448 = queryNorm
            0.15067379 = fieldWeight in 3650, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4093587 = idf(docFreq=3973, maxDocs=44218)
              0.03125 = fieldNorm(doc=3650)
        0.008966677 = product of:
          0.017933354 = sum of:
            0.017933354 = weight(_text_:ed in 3650) [ClassicSimilarity], result of:
              0.017933354 = score(doc=3650,freq=2.0), product of:
                0.11411327 = queryWeight, product of:
                  3.5559888 = idf(docFreq=3431, maxDocs=44218)
                  0.032090448 = queryNorm
                0.15715398 = fieldWeight in 3650, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5559888 = idf(docFreq=3431, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3650)
          0.5 = coord(1/2)
      0.125 = coord(2/16)
    
    Footnote
    Original in: Journal of the American Society for Information Science 21(1970) no.3, S.187-194.
    Source
    Theory of subject analysis: a sourcebook. Ed.: L.M. Chan, et al
  3. Witschel, H.F.: Terminology extraction and automatic indexing : comparison and qualitative evaluation of methods (2005) 0.00
    0.0025791824 = product of:
      0.04126692 = sum of:
        0.04126692 = weight(_text_:author in 1842) [ClassicSimilarity], result of:
          0.04126692 = score(doc=1842,freq=2.0), product of:
            0.15482868 = queryWeight, product of:
              4.824759 = idf(docFreq=964, maxDocs=44218)
              0.032090448 = queryNorm
            0.26653278 = fieldWeight in 1842, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.824759 = idf(docFreq=964, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1842)
      0.0625 = coord(1/16)
    
    Abstract
    Many terminology engineering processes involve the task of automatic terminology extraction: before the terminology of a given domain can be modelled, organised or standardised, important concepts (or terms) of this domain have to be identified and fed into terminological databases. These serve in further steps as a starting point for compiling dictionaries, thesauri or maybe even terminological ontologies for the domain. For the extraction of the initial concepts, extraction methods are needed that operate on specialised language texts. On the other hand, many machine learning or information retrieval applications require automatic indexing techniques. In Machine Learning applications concerned with the automatic clustering or classification of texts, often feature vectors are needed that describe the contents of a given text briefly but meaningfully. These feature vectors typically consist of a fairly small set of index terms together with weights indicating their importance. Short but meaningful descriptions of document contents as provided by good index terms are also useful to humans: some knowledge management applications (e.g. topic maps) use them as a set of basic concepts (topics). The author believes that the tasks of terminology extraction and automatic indexing have much in common and can thus benefit from the same set of basic algorithms. It is the goal of this paper to outline some methods that may be used in both contexts, but also to find the discriminating factors between the two tasks that call for the variation of parameters or application of different techniques. The discussion of these methods will be based on statistical, syntactical and especially morphological properties of (index) terms. The paper is concluded by the presentation of some qualitative and quantitative results comparing statistical and morphological methods.
  4. Porter, M.F.: ¬An algorithm for suffix stripping (1980) 0.00
    0.001681252 = product of:
      0.026900033 = sum of:
        0.026900033 = product of:
          0.053800065 = sum of:
            0.053800065 = weight(_text_:ed in 3122) [ClassicSimilarity], result of:
              0.053800065 = score(doc=3122,freq=2.0), product of:
                0.11411327 = queryWeight, product of:
                  3.5559888 = idf(docFreq=3431, maxDocs=44218)
                  0.032090448 = queryNorm
                0.47146195 = fieldWeight in 3122, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5559888 = idf(docFreq=3431, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3122)
          0.5 = coord(1/2)
      0.0625 = coord(1/16)
    
    Footnote
    Wiederabgedruckt in: Readings in information retrieval. Ed.: K. Sparck Jones u. P. Willett. San Francisco: Morgan Kaufmann 1997. S.313-316.
  5. Goller, C.; Löning, J.; Will, T.; Wolff, W.: Automatic document classification : a thourough evaluation of various methods (2000) 0.00
    0.0016581976 = product of:
      0.026531162 = sum of:
        0.026531162 = weight(_text_:26 in 5480) [ClassicSimilarity], result of:
          0.026531162 = score(doc=5480,freq=2.0), product of:
            0.113328174 = queryWeight, product of:
              3.5315237 = idf(docFreq=3516, maxDocs=44218)
              0.032090448 = queryNorm
            0.23410915 = fieldWeight in 5480, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5315237 = idf(docFreq=3516, maxDocs=44218)
              0.046875 = fieldNorm(doc=5480)
      0.0625 = coord(1/16)
    
    Date
    17. 1.1997 17:19:26
  6. Li, W.; Wong, K.-F.; Yuan, C.: Toward automatic Chinese temporal information extraction (2001) 0.00
    0.0012878824 = product of:
      0.02060612 = sum of:
        0.02060612 = weight(_text_:american in 6029) [ClassicSimilarity], result of:
          0.02060612 = score(doc=6029,freq=2.0), product of:
            0.10940785 = queryWeight, product of:
              3.4093587 = idf(docFreq=3973, maxDocs=44218)
              0.032090448 = queryNorm
            0.18834224 = fieldWeight in 6029, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4093587 = idf(docFreq=3973, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6029)
      0.0625 = coord(1/16)
    
    Source
    Journal of the American Society for Information Science and technology. 52(2001) no.9, S.748-762
  7. Fagan, J.L.: ¬The effectiveness of a nonsyntactic approach to automatic phrase indexing for document retrieval (1989) 0.00
    0.0012878824 = product of:
      0.02060612 = sum of:
        0.02060612 = weight(_text_:american in 1845) [ClassicSimilarity], result of:
          0.02060612 = score(doc=1845,freq=2.0), product of:
            0.10940785 = queryWeight, product of:
              3.4093587 = idf(docFreq=3973, maxDocs=44218)
              0.032090448 = queryNorm
            0.18834224 = fieldWeight in 1845, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4093587 = idf(docFreq=3973, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1845)
      0.0625 = coord(1/16)
    
    Source
    Journal of the American Society for Information Science. 40(1989) no.2, S.115-132
  8. Fox, C.: Lexical analysis and stoplists (1992) 0.00
    0.0011208346 = product of:
      0.017933354 = sum of:
        0.017933354 = product of:
          0.035866708 = sum of:
            0.035866708 = weight(_text_:ed in 3502) [ClassicSimilarity], result of:
              0.035866708 = score(doc=3502,freq=2.0), product of:
                0.11411327 = queryWeight, product of:
                  3.5559888 = idf(docFreq=3431, maxDocs=44218)
                  0.032090448 = queryNorm
                0.31430796 = fieldWeight in 3502, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5559888 = idf(docFreq=3431, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3502)
          0.5 = coord(1/2)
      0.0625 = coord(1/16)
    
    Source
    Information retrieval: data structures and algorithms. Ed.: W.B. Frakes u. R. Baeza-Yates
  9. Riloff, E.: ¬An empirical study of automated dictionary construction for information extraction in three domains (1996) 0.00
    0.0010869532 = product of:
      0.017391251 = sum of:
        0.017391251 = product of:
          0.034782503 = sum of:
            0.034782503 = weight(_text_:22 in 6752) [ClassicSimilarity], result of:
              0.034782503 = score(doc=6752,freq=2.0), product of:
                0.11237528 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.032090448 = queryNorm
                0.30952093 = fieldWeight in 6752, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6752)
          0.5 = coord(1/2)
      0.0625 = coord(1/16)
    
    Date
    6. 3.1997 16:22:15
  10. Wacholder, N.; Byrd, R.J.: Retrieving information from full text using linguistic knowledge (1994) 0.00
    8.40626E-4 = product of:
      0.013450016 = sum of:
        0.013450016 = product of:
          0.026900033 = sum of:
            0.026900033 = weight(_text_:ed in 8524) [ClassicSimilarity], result of:
              0.026900033 = score(doc=8524,freq=2.0), product of:
                0.11411327 = queryWeight, product of:
                  3.5559888 = idf(docFreq=3431, maxDocs=44218)
                  0.032090448 = queryNorm
                0.23573098 = fieldWeight in 8524, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5559888 = idf(docFreq=3431, maxDocs=44218)
                  0.046875 = fieldNorm(doc=8524)
          0.5 = coord(1/2)
      0.0625 = coord(1/16)
    
    Source
    Proceedings of the 15th National Online Meeting 1994, New York, 10-12 May 1994. Ed. by M.E. Williams
  11. Needham, R.M.; Sparck Jones, K.: Keywords and clumps (1985) 0.00
    4.903652E-4 = product of:
      0.007845843 = sum of:
        0.007845843 = product of:
          0.015691686 = sum of:
            0.015691686 = weight(_text_:ed in 3645) [ClassicSimilarity], result of:
              0.015691686 = score(doc=3645,freq=2.0), product of:
                0.11411327 = queryWeight, product of:
                  3.5559888 = idf(docFreq=3431, maxDocs=44218)
                  0.032090448 = queryNorm
                0.13750973 = fieldWeight in 3645, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5559888 = idf(docFreq=3431, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=3645)
          0.5 = coord(1/2)
      0.0625 = coord(1/16)
    
    Source
    Theory of subject analysis: a sourcebook. Ed.: L.M. Chan, et al