Search (5 results, page 1 of 1)

  • × theme_ss:"Automatisches Klassifizieren"
  • × year_i:[2010 TO 2020}
  1. HaCohen-Kerner, Y. et al.: Classification using various machine learning methods and combinations of key-phrases and visual features (2016) 0.02
    0.017346188 = product of:
      0.034692377 = sum of:
        0.034692377 = product of:
          0.06938475 = sum of:
            0.06938475 = weight(_text_:22 in 2748) [ClassicSimilarity], result of:
              0.06938475 = score(doc=2748,freq=2.0), product of:
                0.17933457 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051211677 = queryNorm
                0.38690117 = fieldWeight in 2748, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2748)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    1. 2.2016 18:25:22
  2. Wartena, C.; Sommer, M.: Automatic classification of scientific records using the German Subject Heading Authority File (SWD) (2012) 0.02
    0.01663633 = product of:
      0.03327266 = sum of:
        0.03327266 = product of:
          0.06654532 = sum of:
            0.06654532 = weight(_text_:headings in 472) [ClassicSimilarity], result of:
              0.06654532 = score(doc=472,freq=2.0), product of:
                0.24837378 = queryWeight, product of:
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.051211677 = queryNorm
                0.2679241 = fieldWeight in 472, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.849944 = idf(docFreq=940, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=472)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The following paper deals with an automatic text classification method which does not require training documents. For this method the German Subject Heading Authority File (SWD), provided by the linked data service of the German National Library is used. Recently the SWD was enriched with notations of the Dewey Decimal Classification (DDC). In consequence it became possible to utilize the subject headings as textual representations for the notations of the DDC. Basically, we we derive the classification of a text from the classification of the words in the text given by the thesaurus. The method was tested by classifying 3826 OAI-Records from 7 different repositories. Mean reciprocal rank and recall were chosen as evaluation measure. Direct comparison to a machine learning method has shown that this method is definitely competitive. Thus we can conclude that the enriched version of the SWD provides high quality information with a broad coverage for classification of German scientific articles.
  3. Zhu, W.Z.; Allen, R.B.: Document clustering using the LSI subspace signature model (2013) 0.01
    0.010407712 = product of:
      0.020815425 = sum of:
        0.020815425 = product of:
          0.04163085 = sum of:
            0.04163085 = weight(_text_:22 in 690) [ClassicSimilarity], result of:
              0.04163085 = score(doc=690,freq=2.0), product of:
                0.17933457 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051211677 = queryNorm
                0.23214069 = fieldWeight in 690, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=690)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    23. 3.2013 13:22:36
  4. Egbert, J.; Biber, D.; Davies, M.: Developing a bottom-up, user-based method of web register classification (2015) 0.01
    0.010407712 = product of:
      0.020815425 = sum of:
        0.020815425 = product of:
          0.04163085 = sum of:
            0.04163085 = weight(_text_:22 in 2158) [ClassicSimilarity], result of:
              0.04163085 = score(doc=2158,freq=2.0), product of:
                0.17933457 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051211677 = queryNorm
                0.23214069 = fieldWeight in 2158, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2158)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    4. 8.2015 19:22:04
  5. Liu, R.-L.: ¬A passage extractor for classification of disease aspect information (2013) 0.01
    0.008673094 = product of:
      0.017346188 = sum of:
        0.017346188 = product of:
          0.034692377 = sum of:
            0.034692377 = weight(_text_:22 in 1107) [ClassicSimilarity], result of:
              0.034692377 = score(doc=1107,freq=2.0), product of:
                0.17933457 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051211677 = queryNorm
                0.19345059 = fieldWeight in 1107, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1107)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    28.10.2013 19:22:57