Search (4 results, page 1 of 1)

  • × theme_ss:"Automatisches Klassifizieren"
  • × theme_ss:"Computerlinguistik"
  1. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.05
    0.048444517 = product of:
      0.07266677 = sum of:
        0.054179683 = product of:
          0.21671873 = sum of:
            0.21671873 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
              0.21671873 = score(doc=562,freq=2.0), product of:
                0.38560814 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.045483325 = queryNorm
                0.56201804 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.25 = coord(1/4)
        0.018487087 = product of:
          0.036974173 = sum of:
            0.036974173 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
              0.036974173 = score(doc=562,freq=2.0), product of:
                0.15927485 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045483325 = queryNorm
                0.23214069 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  2. Frobese, D.T.: Klassifikationsaufgaben mit der SENTRAX : Konkreter Fall: Automatische Detektion von SPAM (2006) 0.01
    0.0082606 = product of:
      0.024781799 = sum of:
        0.024781799 = product of:
          0.049563598 = sum of:
            0.049563598 = weight(_text_:von in 5980) [ClassicSimilarity], result of:
              0.049563598 = score(doc=5980,freq=6.0), product of:
                0.12134718 = queryWeight, product of:
                  2.6679487 = idf(docFreq=8340, maxDocs=44218)
                  0.045483325 = queryNorm
                0.40844458 = fieldWeight in 5980, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  2.6679487 = idf(docFreq=8340, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5980)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Die Suchfunktionen des SENTRAX-Verfahrens werden für die Klassifizierung von Mails und im Besonderen für die Detektion von SPAM eingesetzt. Die Eigenschaften einer kontextähnlichen Suche und die Fehlertoleranz sollen genutzt werden, um SPAM Nachrichten treffsicher aufzuspüren.
  3. Ruiz, M.E.; Srinivasan, P.: Combining machine learning and hierarchical indexing structures for text categorization (2001) 0.01
    0.007237594 = product of:
      0.021712782 = sum of:
        0.021712782 = weight(_text_:h in 1595) [ClassicSimilarity], result of:
          0.021712782 = score(doc=1595,freq=2.0), product of:
            0.113001004 = queryWeight, product of:
              2.4844491 = idf(docFreq=10020, maxDocs=44218)
              0.045483325 = queryNorm
            0.19214681 = fieldWeight in 1595, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.4844491 = idf(docFreq=10020, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1595)
      0.33333334 = coord(1/3)
    
    Source
    Advances in classification research, vol.10: proceedings of the 10th ASIS SIG/CR Classification Research Workshop. Ed.: Albrechtsen, H. u. J.E. Mai
  4. Zhang, X: Rough set theory based automatic text categorization (2005) 0.01
    0.0067447512 = product of:
      0.020234253 = sum of:
        0.020234253 = product of:
          0.040468507 = sum of:
            0.040468507 = weight(_text_:von in 2822) [ClassicSimilarity], result of:
              0.040468507 = score(doc=2822,freq=4.0), product of:
                0.12134718 = queryWeight, product of:
                  2.6679487 = idf(docFreq=8340, maxDocs=44218)
                  0.045483325 = queryNorm
                0.3334936 = fieldWeight in 2822, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.6679487 = idf(docFreq=8340, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2822)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Der Forschungsbericht "Rough Set Theory Based Automatic Text Categorization and the Handling of Semantic Heterogeneity" von Xueying Zhang ist in Buchform auf Englisch erschienen. Zhang hat in ihrer Arbeit ein Verfahren basierend auf der Rough Set Theory entwickelt, das Beziehungen zwischen Schlagwörtern verschiedener Vokabulare herstellt. Sie war von 2003 bis 2005 Mitarbeiterin des IZ und ist seit Oktober 2005 Associate Professor an der Nanjing University of Science and Technology.

Languages

Types