Search (21 results, page 1 of 2)

  • × theme_ss:"Automatisches Klassifizieren"
  1. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.32
    0.3182387 = sum of:
      0.07477565 = product of:
        0.22432694 = sum of:
          0.22432694 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
            0.22432694 = score(doc=562,freq=2.0), product of:
              0.39914545 = queryWeight, product of:
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.04708008 = queryNorm
              0.56201804 = fieldWeight in 562, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.046875 = fieldNorm(doc=562)
        0.33333334 = coord(1/3)
      0.22432694 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
        0.22432694 = score(doc=562,freq=2.0), product of:
          0.39914545 = queryWeight, product of:
            8.478011 = idf(docFreq=24, maxDocs=44218)
            0.04708008 = queryNorm
          0.56201804 = fieldWeight in 562, product of:
            1.4142135 = tf(freq=2.0), with freq of:
              2.0 = termFreq=2.0
            8.478011 = idf(docFreq=24, maxDocs=44218)
            0.046875 = fieldNorm(doc=562)
      0.019136103 = product of:
        0.038272206 = sum of:
          0.038272206 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
            0.038272206 = score(doc=562,freq=2.0), product of:
              0.16486642 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04708008 = queryNorm
              0.23214069 = fieldWeight in 562, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=562)
        0.5 = coord(1/2)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  2. Subramanian, S.; Shafer, K.E.: Clustering (2001) 0.01
    0.012757402 = product of:
      0.038272206 = sum of:
        0.038272206 = product of:
          0.07654441 = sum of:
            0.07654441 = weight(_text_:22 in 1046) [ClassicSimilarity], result of:
              0.07654441 = score(doc=1046,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.46428138 = fieldWeight in 1046, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=1046)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    5. 5.2003 14:17:22
  3. Golub, K.: Automated subject classification of textual Web pages, based on a controlled vocabulary : challenges and recommendations (2006) 0.01
    0.011185795 = product of:
      0.033557385 = sum of:
        0.033557385 = product of:
          0.06711477 = sum of:
            0.06711477 = weight(_text_:multimedia in 5897) [ClassicSimilarity], result of:
              0.06711477 = score(doc=5897,freq=2.0), product of:
                0.21832302 = queryWeight, product of:
                  4.6372695 = idf(docFreq=1163, maxDocs=44218)
                  0.04708008 = queryNorm
                0.30741042 = fieldWeight in 5897, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.6372695 = idf(docFreq=1163, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5897)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Source
    New review of hypermedia and multimedia. 12(2006) no.1, S.11-27
  4. Denoyer, L.; Gallinari, P.: Bayesian network model for semi-structured document classification (2004) 0.01
    0.011185795 = product of:
      0.033557385 = sum of:
        0.033557385 = product of:
          0.06711477 = sum of:
            0.06711477 = weight(_text_:multimedia in 995) [ClassicSimilarity], result of:
              0.06711477 = score(doc=995,freq=2.0), product of:
                0.21832302 = queryWeight, product of:
                  4.6372695 = idf(docFreq=1163, maxDocs=44218)
                  0.04708008 = queryNorm
                0.30741042 = fieldWeight in 995, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.6372695 = idf(docFreq=1163, maxDocs=44218)
                  0.046875 = fieldNorm(doc=995)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Recently, a new community has started to emerge around the development of new information research methods for searching and analyzing semi-structured and XML like documents. The goal is to handle both content and structural information, and to deal with different types of information content (text, image, etc.). We consider here the task of structured document classification. We propose a generative model able to handle both structure and content which is based on Bayesian networks. We then show how to transform this generative model into a discriminant classifier using the method of Fisher kernel. The model is then extended for dealing with different types of content information (here text and images). The model was tested on three databases: the classical webKB corpus composed of HTML pages, the new INEX corpus which has become a reference in the field of ad-hoc retrieval for XML documents, and a multimedia corpus of Web pages.
  5. Reiner, U.: Automatische DDC-Klassifizierung von bibliografischen Titeldatensätzen (2009) 0.01
    0.010631169 = product of:
      0.031893507 = sum of:
        0.031893507 = product of:
          0.06378701 = sum of:
            0.06378701 = weight(_text_:22 in 611) [ClassicSimilarity], result of:
              0.06378701 = score(doc=611,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.38690117 = fieldWeight in 611, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=611)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 8.2009 12:54:24
  6. HaCohen-Kerner, Y. et al.: Classification using various machine learning methods and combinations of key-phrases and visual features (2016) 0.01
    0.010631169 = product of:
      0.031893507 = sum of:
        0.031893507 = product of:
          0.06378701 = sum of:
            0.06378701 = weight(_text_:22 in 2748) [ClassicSimilarity], result of:
              0.06378701 = score(doc=2748,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.38690117 = fieldWeight in 2748, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2748)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    1. 2.2016 18:25:22
  7. Bock, H.-H.: Datenanalyse zur Strukturierung und Ordnung von Information (1989) 0.01
    0.007441818 = product of:
      0.022325452 = sum of:
        0.022325452 = product of:
          0.044650905 = sum of:
            0.044650905 = weight(_text_:22 in 141) [ClassicSimilarity], result of:
              0.044650905 = score(doc=141,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.2708308 = fieldWeight in 141, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=141)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Pages
    S.1-22
  8. Dubin, D.: Dimensions and discriminability (1998) 0.01
    0.007441818 = product of:
      0.022325452 = sum of:
        0.022325452 = product of:
          0.044650905 = sum of:
            0.044650905 = weight(_text_:22 in 2338) [ClassicSimilarity], result of:
              0.044650905 = score(doc=2338,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.2708308 = fieldWeight in 2338, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2338)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 9.1997 19:16:05
  9. Automatic classification research at OCLC (2002) 0.01
    0.007441818 = product of:
      0.022325452 = sum of:
        0.022325452 = product of:
          0.044650905 = sum of:
            0.044650905 = weight(_text_:22 in 1563) [ClassicSimilarity], result of:
              0.044650905 = score(doc=1563,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.2708308 = fieldWeight in 1563, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1563)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    5. 5.2003 9:22:09
  10. Jenkins, C.: Automatic classification of Web resources using Java and Dewey Decimal Classification (1998) 0.01
    0.007441818 = product of:
      0.022325452 = sum of:
        0.022325452 = product of:
          0.044650905 = sum of:
            0.044650905 = weight(_text_:22 in 1673) [ClassicSimilarity], result of:
              0.044650905 = score(doc=1673,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.2708308 = fieldWeight in 1673, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1673)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    1. 8.1996 22:08:06
  11. Yoon, Y.; Lee, C.; Lee, G.G.: ¬An effective procedure for constructing a hierarchical text classification system (2006) 0.01
    0.007441818 = product of:
      0.022325452 = sum of:
        0.022325452 = product of:
          0.044650905 = sum of:
            0.044650905 = weight(_text_:22 in 5273) [ClassicSimilarity], result of:
              0.044650905 = score(doc=5273,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.2708308 = fieldWeight in 5273, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5273)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 7.2006 16:24:52
  12. Yi, K.: Automatic text classification using library classification schemes : trends, issues and challenges (2007) 0.01
    0.007441818 = product of:
      0.022325452 = sum of:
        0.022325452 = product of:
          0.044650905 = sum of:
            0.044650905 = weight(_text_:22 in 2560) [ClassicSimilarity], result of:
              0.044650905 = score(doc=2560,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.2708308 = fieldWeight in 2560, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2560)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 9.2008 18:31:54
  13. Liu, R.-L.: Context recognition for hierarchical text classification (2009) 0.01
    0.006378701 = product of:
      0.019136103 = sum of:
        0.019136103 = product of:
          0.038272206 = sum of:
            0.038272206 = weight(_text_:22 in 2760) [ClassicSimilarity], result of:
              0.038272206 = score(doc=2760,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.23214069 = fieldWeight in 2760, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2760)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 3.2009 19:11:54
  14. Pfeffer, M.: Automatische Vergabe von RVK-Notationen mittels fallbasiertem Schließen (2009) 0.01
    0.006378701 = product of:
      0.019136103 = sum of:
        0.019136103 = product of:
          0.038272206 = sum of:
            0.038272206 = weight(_text_:22 in 3051) [ClassicSimilarity], result of:
              0.038272206 = score(doc=3051,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.23214069 = fieldWeight in 3051, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3051)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 8.2009 19:51:28
  15. Zhu, W.Z.; Allen, R.B.: Document clustering using the LSI subspace signature model (2013) 0.01
    0.006378701 = product of:
      0.019136103 = sum of:
        0.019136103 = product of:
          0.038272206 = sum of:
            0.038272206 = weight(_text_:22 in 690) [ClassicSimilarity], result of:
              0.038272206 = score(doc=690,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.23214069 = fieldWeight in 690, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=690)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    23. 3.2013 13:22:36
  16. Egbert, J.; Biber, D.; Davies, M.: Developing a bottom-up, user-based method of web register classification (2015) 0.01
    0.006378701 = product of:
      0.019136103 = sum of:
        0.019136103 = product of:
          0.038272206 = sum of:
            0.038272206 = weight(_text_:22 in 2158) [ClassicSimilarity], result of:
              0.038272206 = score(doc=2158,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.23214069 = fieldWeight in 2158, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2158)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    4. 8.2015 19:22:04
  17. Dolin, R.; Agrawal, D.; El Abbadi, A.; Pearlman, J.: Using automated classification for summarizing and selecting heterogeneous information sources (1998) 0.01
    0.0055928975 = product of:
      0.016778693 = sum of:
        0.016778693 = product of:
          0.033557385 = sum of:
            0.033557385 = weight(_text_:multimedia in 1253) [ClassicSimilarity], result of:
              0.033557385 = score(doc=1253,freq=2.0), product of:
                0.21832302 = queryWeight, product of:
                  4.6372695 = idf(docFreq=1163, maxDocs=44218)
                  0.04708008 = queryNorm
                0.15370521 = fieldWeight in 1253, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.6372695 = idf(docFreq=1163, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1253)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Information retrieval over the Internet increasingly requires the filtering of thousands of heterogeneous information sources. Important sources of information include not only traditional databases with structured data and queries, but also increasing numbers of non-traditional, semi- or unstructured collections such as Web sites, FTP archives, etc. As the number and variability of sources increases, new ways of automatically summarizing, discovering, and selecting collections relevant to a user's query are needed. One such method involves the use of classification schemes, such as the Library of Congress Classification (LCC), within which a collection may be represented based on its content, irrespective of the structure of the actual data or documents. For such a system to be useful in a large-scale distributed environment, it must be easy to use for both collection managers and users. As a result, it must be possible to classify documents automatically within a classification scheme. Furthermore, there must be a straightforward and intuitive interface with which the user may use the scheme to assist in information retrieval (IR). Our work with the Alexandria Digital Library (ADL) Project focuses on geo-referenced information, whether text, maps, aerial photographs, or satellite images. As a result, we have emphasized techniques which work with both text and non-text, such as combined textual and graphical queries, multi-dimensional indexing, and IR methods which are not solely dependent on words or phrases. Part of this work involves locating relevant online sources of information. In particular, we have designed and are currently testing aspects of an architecture, Pharos, which we believe will scale up to 1.000.000 heterogeneous sources. Pharos accommodates heterogeneity in content and format, both among multiple sources as well as within a single source. That is, we consider sources to include Web sites, FTP archives, newsgroups, and full digital libraries; all of these systems can include a wide variety of content and multimedia data formats. Pharos is based on the use of hierarchical classification schemes. These include not only well-known 'subject' (or 'concept') based schemes such as the Dewey Decimal System and the LCC, but also, for example, geographic classifications, which might be constructed as layers of smaller and smaller hierarchical longitude/latitude boxes. Pharos is designed to work with sophisticated queries which utilize subjects, geographical locations, temporal specifications, and other types of information domains. The Pharos architecture requires that hierarchically structured collection metadata be extracted so that it can be partitioned in such a way as to greatly enhance scalability. Automated classification is important to Pharos because it allows information sources to extract the requisite collection metadata automatically that must be distributed.
  18. Mengle, S.; Goharian, N.: Passage detection using text classification (2009) 0.01
    0.0053155846 = product of:
      0.015946753 = sum of:
        0.015946753 = product of:
          0.031893507 = sum of:
            0.031893507 = weight(_text_:22 in 2765) [ClassicSimilarity], result of:
              0.031893507 = score(doc=2765,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.19345059 = fieldWeight in 2765, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2765)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 3.2009 19:14:43
  19. Liu, R.-L.: ¬A passage extractor for classification of disease aspect information (2013) 0.01
    0.0053155846 = product of:
      0.015946753 = sum of:
        0.015946753 = product of:
          0.031893507 = sum of:
            0.031893507 = weight(_text_:22 in 1107) [ClassicSimilarity], result of:
              0.031893507 = score(doc=1107,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.19345059 = fieldWeight in 1107, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1107)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    28.10.2013 19:22:57
  20. Khoo, C.S.G.; Ng, K.; Ou, S.: ¬An exploratory study of human clustering of Web pages (2003) 0.00
    0.0042524673 = product of:
      0.012757402 = sum of:
        0.012757402 = product of:
          0.025514804 = sum of:
            0.025514804 = weight(_text_:22 in 2741) [ClassicSimilarity], result of:
              0.025514804 = score(doc=2741,freq=2.0), product of:
                0.16486642 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04708008 = queryNorm
                0.15476047 = fieldWeight in 2741, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2741)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    12. 9.2004 9:56:22