Search (20 results, page 1 of 1)

  • × theme_ss:"Automatisches Klassifizieren"
  1. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.10
    0.09965296 = sum of:
      0.07934699 = product of:
        0.23804097 = sum of:
          0.23804097 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
            0.23804097 = score(doc=562,freq=2.0), product of:
              0.42354685 = queryWeight, product of:
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.04995828 = queryNorm
              0.56201804 = fieldWeight in 562, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.046875 = fieldNorm(doc=562)
        0.33333334 = coord(1/3)
      0.02030597 = product of:
        0.04061194 = sum of:
          0.04061194 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
            0.04061194 = score(doc=562,freq=2.0), product of:
              0.17494538 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04995828 = queryNorm
              0.23214069 = fieldWeight in 562, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=562)
        0.5 = coord(1/2)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  2. Giorgetti, D.; Sebastiani, F.: Automating survey coding by multiclass text categorization techniques (2003) 0.03
    0.029595422 = product of:
      0.059190843 = sum of:
        0.059190843 = product of:
          0.11838169 = sum of:
            0.11838169 = weight(_text_:opinion in 5172) [ClassicSimilarity], result of:
              0.11838169 = score(doc=5172,freq=2.0), product of:
                0.3271964 = queryWeight, product of:
                  6.5493927 = idf(docFreq=171, maxDocs=44218)
                  0.04995828 = queryNorm
                0.3618062 = fieldWeight in 5172, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  6.5493927 = idf(docFreq=171, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5172)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    In this issue Giorgetti, and Sebastiani suggest that answers to open ended questions in survey instruments can be coded automatically by creating classifiers which learn from training sets of manually coded answers. The manual effort required is only that of classifying a representative set of documents, not creating a dictionary of words that trigger an assignment. They use a naive Bayesian probabilistic learner from Mc Callum's RAINBOW package and the multi-class support vector machine learner from Hsu and Lin's BSVM package, both examples of text categorization techniques. Data from the 1996 General Social Survey by the U.S. National Opinion Research Center provided a set of answers to three questions (previously tested by Viechnicki using a dictionary approach), their associated manually assigned category codes, and a complete set of predefined category codes. The learners were run on three random disjoint subsets of the answer sets to create the classifiers and a remaining set was used as a test set. The dictionary approach is out preformed by 18% for RAINBOW and by 17% for BSVM, while the standard deviation of the results is reduced by 28% and 34% respectively over the dictionary approach.
  3. Vilares, D.; Alonso, M.A.; Gómez-Rodríguez, C.: On the usefulness of lexical and syntactic processing in polarity classification of Twitter messages (2015) 0.03
    0.029595422 = product of:
      0.059190843 = sum of:
        0.059190843 = product of:
          0.11838169 = sum of:
            0.11838169 = weight(_text_:opinion in 2161) [ClassicSimilarity], result of:
              0.11838169 = score(doc=2161,freq=2.0), product of:
                0.3271964 = queryWeight, product of:
                  6.5493927 = idf(docFreq=171, maxDocs=44218)
                  0.04995828 = queryNorm
                0.3618062 = fieldWeight in 2161, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  6.5493927 = idf(docFreq=171, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2161)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Millions of micro texts are published every day on Twitter. Identifying the sentiment present in them can be helpful for measuring the frame of mind of the public, their satisfaction with respect to a product, or their support of a social event. In this context, polarity classification is a subfield of sentiment analysis focused on determining whether the content of a text is objective or subjective, and in the latter case, if it conveys a positive or a negative opinion. Most polarity detection techniques tend to take into account individual terms in the text and even some degree of linguistic knowledge, but they do not usually consider syntactic relations between words. This article explores how relating lexical, syntactic, and psychometric information can be helpful to perform polarity classification on Spanish tweets. We provide an evaluation for both shallow and deep linguistic perspectives. Empirical results show an improved performance of syntactic approaches over pure lexical models when using large training sets to create a classifier, but this tendency is reversed when small training collections are used.
  4. Subramanian, S.; Shafer, K.E.: Clustering (2001) 0.02
    0.02030597 = product of:
      0.04061194 = sum of:
        0.04061194 = product of:
          0.08122388 = sum of:
            0.08122388 = weight(_text_:22 in 1046) [ClassicSimilarity], result of:
              0.08122388 = score(doc=1046,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.46428138 = fieldWeight in 1046, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=1046)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    5. 5.2003 14:17:22
  5. Reiner, U.: Automatische DDC-Klassifizierung von bibliografischen Titeldatensätzen (2009) 0.02
    0.016921643 = product of:
      0.033843286 = sum of:
        0.033843286 = product of:
          0.06768657 = sum of:
            0.06768657 = weight(_text_:22 in 611) [ClassicSimilarity], result of:
              0.06768657 = score(doc=611,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.38690117 = fieldWeight in 611, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=611)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 8.2009 12:54:24
  6. HaCohen-Kerner, Y. et al.: Classification using various machine learning methods and combinations of key-phrases and visual features (2016) 0.02
    0.016921643 = product of:
      0.033843286 = sum of:
        0.033843286 = product of:
          0.06768657 = sum of:
            0.06768657 = weight(_text_:22 in 2748) [ClassicSimilarity], result of:
              0.06768657 = score(doc=2748,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.38690117 = fieldWeight in 2748, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2748)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    1. 2.2016 18:25:22
  7. Bock, H.-H.: Datenanalyse zur Strukturierung und Ordnung von Information (1989) 0.01
    0.01184515 = product of:
      0.0236903 = sum of:
        0.0236903 = product of:
          0.0473806 = sum of:
            0.0473806 = weight(_text_:22 in 141) [ClassicSimilarity], result of:
              0.0473806 = score(doc=141,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.2708308 = fieldWeight in 141, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=141)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Pages
    S.1-22
  8. Dubin, D.: Dimensions and discriminability (1998) 0.01
    0.01184515 = product of:
      0.0236903 = sum of:
        0.0236903 = product of:
          0.0473806 = sum of:
            0.0473806 = weight(_text_:22 in 2338) [ClassicSimilarity], result of:
              0.0473806 = score(doc=2338,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.2708308 = fieldWeight in 2338, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2338)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 9.1997 19:16:05
  9. Automatic classification research at OCLC (2002) 0.01
    0.01184515 = product of:
      0.0236903 = sum of:
        0.0236903 = product of:
          0.0473806 = sum of:
            0.0473806 = weight(_text_:22 in 1563) [ClassicSimilarity], result of:
              0.0473806 = score(doc=1563,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.2708308 = fieldWeight in 1563, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1563)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    5. 5.2003 9:22:09
  10. Jenkins, C.: Automatic classification of Web resources using Java and Dewey Decimal Classification (1998) 0.01
    0.01184515 = product of:
      0.0236903 = sum of:
        0.0236903 = product of:
          0.0473806 = sum of:
            0.0473806 = weight(_text_:22 in 1673) [ClassicSimilarity], result of:
              0.0473806 = score(doc=1673,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.2708308 = fieldWeight in 1673, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1673)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    1. 8.1996 22:08:06
  11. Yoon, Y.; Lee, C.; Lee, G.G.: ¬An effective procedure for constructing a hierarchical text classification system (2006) 0.01
    0.01184515 = product of:
      0.0236903 = sum of:
        0.0236903 = product of:
          0.0473806 = sum of:
            0.0473806 = weight(_text_:22 in 5273) [ClassicSimilarity], result of:
              0.0473806 = score(doc=5273,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.2708308 = fieldWeight in 5273, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5273)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 7.2006 16:24:52
  12. Yi, K.: Automatic text classification using library classification schemes : trends, issues and challenges (2007) 0.01
    0.01184515 = product of:
      0.0236903 = sum of:
        0.0236903 = product of:
          0.0473806 = sum of:
            0.0473806 = weight(_text_:22 in 2560) [ClassicSimilarity], result of:
              0.0473806 = score(doc=2560,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.2708308 = fieldWeight in 2560, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2560)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 9.2008 18:31:54
  13. Liu, R.-L.: Context recognition for hierarchical text classification (2009) 0.01
    0.010152985 = product of:
      0.02030597 = sum of:
        0.02030597 = product of:
          0.04061194 = sum of:
            0.04061194 = weight(_text_:22 in 2760) [ClassicSimilarity], result of:
              0.04061194 = score(doc=2760,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.23214069 = fieldWeight in 2760, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2760)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 3.2009 19:11:54
  14. Pfeffer, M.: Automatische Vergabe von RVK-Notationen mittels fallbasiertem Schließen (2009) 0.01
    0.010152985 = product of:
      0.02030597 = sum of:
        0.02030597 = product of:
          0.04061194 = sum of:
            0.04061194 = weight(_text_:22 in 3051) [ClassicSimilarity], result of:
              0.04061194 = score(doc=3051,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.23214069 = fieldWeight in 3051, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3051)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 8.2009 19:51:28
  15. Zhu, W.Z.; Allen, R.B.: Document clustering using the LSI subspace signature model (2013) 0.01
    0.010152985 = product of:
      0.02030597 = sum of:
        0.02030597 = product of:
          0.04061194 = sum of:
            0.04061194 = weight(_text_:22 in 690) [ClassicSimilarity], result of:
              0.04061194 = score(doc=690,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.23214069 = fieldWeight in 690, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=690)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    23. 3.2013 13:22:36
  16. Egbert, J.; Biber, D.; Davies, M.: Developing a bottom-up, user-based method of web register classification (2015) 0.01
    0.010152985 = product of:
      0.02030597 = sum of:
        0.02030597 = product of:
          0.04061194 = sum of:
            0.04061194 = weight(_text_:22 in 2158) [ClassicSimilarity], result of:
              0.04061194 = score(doc=2158,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.23214069 = fieldWeight in 2158, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2158)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    4. 8.2015 19:22:04
  17. Mengle, S.; Goharian, N.: Passage detection using text classification (2009) 0.01
    0.008460822 = product of:
      0.016921643 = sum of:
        0.016921643 = product of:
          0.033843286 = sum of:
            0.033843286 = weight(_text_:22 in 2765) [ClassicSimilarity], result of:
              0.033843286 = score(doc=2765,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.19345059 = fieldWeight in 2765, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2765)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 3.2009 19:14:43
  18. Liu, R.-L.: ¬A passage extractor for classification of disease aspect information (2013) 0.01
    0.008460822 = product of:
      0.016921643 = sum of:
        0.016921643 = product of:
          0.033843286 = sum of:
            0.033843286 = weight(_text_:22 in 1107) [ClassicSimilarity], result of:
              0.033843286 = score(doc=1107,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.19345059 = fieldWeight in 1107, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1107)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    28.10.2013 19:22:57
  19. Khoo, C.S.G.; Ng, K.; Ou, S.: ¬An exploratory study of human clustering of Web pages (2003) 0.01
    0.0067686574 = product of:
      0.013537315 = sum of:
        0.013537315 = product of:
          0.02707463 = sum of:
            0.02707463 = weight(_text_:22 in 2741) [ClassicSimilarity], result of:
              0.02707463 = score(doc=2741,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.15476047 = fieldWeight in 2741, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2741)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    12. 9.2004 9:56:22
  20. Reiner, U.: Automatische DDC-Klassifizierung bibliografischer Titeldatensätze der Deutschen Nationalbibliografie (2009) 0.01
    0.0067686574 = product of:
      0.013537315 = sum of:
        0.013537315 = product of:
          0.02707463 = sum of:
            0.02707463 = weight(_text_:22 in 3284) [ClassicSimilarity], result of:
              0.02707463 = score(doc=3284,freq=2.0), product of:
                0.17494538 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04995828 = queryNorm
                0.15476047 = fieldWeight in 3284, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3284)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 1.2010 14:41:24