Search (2158 results, page 1 of 108)

  • × type_ss:"a"
  • × year_i:[2000 TO 2010}
  1. Mas, S.; Marleau, Y.: Proposition of a faceted classification model to support corporate information organization and digital records management (2009) 0.15
    0.15028802 = product of:
      0.25048003 = sum of:
        0.060036086 = product of:
          0.18010825 = sum of:
            0.18010825 = weight(_text_:3a in 2918) [ClassicSimilarity], result of:
              0.18010825 = score(doc=2918,freq=2.0), product of:
                0.32046703 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.037799787 = queryNorm
                0.56201804 = fieldWeight in 2918, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2918)
          0.33333334 = coord(1/3)
        0.18010825 = weight(_text_:2f in 2918) [ClassicSimilarity], result of:
          0.18010825 = score(doc=2918,freq=2.0), product of:
            0.32046703 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.037799787 = queryNorm
            0.56201804 = fieldWeight in 2918, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=2918)
        0.0103356745 = product of:
          0.031007024 = sum of:
            0.031007024 = weight(_text_:29 in 2918) [ClassicSimilarity], result of:
              0.031007024 = score(doc=2918,freq=2.0), product of:
                0.13296783 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037799787 = queryNorm
                0.23319192 = fieldWeight in 2918, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2918)
          0.33333334 = coord(1/3)
      0.6 = coord(3/5)
    
    Date
    29. 8.2009 21:15:48
    Footnote
    Vgl.: http://ieeexplore.ieee.org/Xplore/login.jsp?reload=true&url=http%3A%2F%2Fieeexplore.ieee.org%2Fiel5%2F4755313%2F4755314%2F04755480.pdf%3Farnumber%3D4755480&authDecision=-203.
  2. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.15
    0.15023223 = product of:
      0.25038704 = sum of:
        0.060036086 = product of:
          0.18010825 = sum of:
            0.18010825 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
              0.18010825 = score(doc=562,freq=2.0), product of:
                0.32046703 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.037799787 = queryNorm
                0.56201804 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
        0.18010825 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.18010825 = score(doc=562,freq=2.0), product of:
            0.32046703 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.037799787 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.010242699 = product of:
          0.030728096 = sum of:
            0.030728096 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
              0.030728096 = score(doc=562,freq=2.0), product of:
                0.13236842 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037799787 = queryNorm
                0.23214069 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
      0.6 = coord(3/5)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  3. Schrodt, R.: Tiefen und Untiefen im wissenschaftlichen Sprachgebrauch (2008) 0.13
    0.12807699 = product of:
      0.32019246 = sum of:
        0.080048114 = product of:
          0.24014434 = sum of:
            0.24014434 = weight(_text_:3a in 140) [ClassicSimilarity], result of:
              0.24014434 = score(doc=140,freq=2.0), product of:
                0.32046703 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.037799787 = queryNorm
                0.7493574 = fieldWeight in 140, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0625 = fieldNorm(doc=140)
          0.33333334 = coord(1/3)
        0.24014434 = weight(_text_:2f in 140) [ClassicSimilarity], result of:
          0.24014434 = score(doc=140,freq=2.0), product of:
            0.32046703 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.037799787 = queryNorm
            0.7493574 = fieldWeight in 140, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0625 = fieldNorm(doc=140)
      0.4 = coord(2/5)
    
    Content
    Vgl. auch: https://studylibde.com/doc/13053640/richard-schrodt. Vgl. auch: http%3A%2F%2Fwww.univie.ac.at%2FGermanistik%2Fschrodt%2Fvorlesung%2Fwissenschaftssprache.doc&usg=AOvVaw1lDLDR6NFf1W0-oC9mEUJf.
  4. Vetere, G.; Lenzerini, M.: Models for semantic interoperability in service-oriented architectures (2005) 0.11
    0.112067364 = product of:
      0.2801684 = sum of:
        0.0700421 = product of:
          0.2101263 = sum of:
            0.2101263 = weight(_text_:3a in 306) [ClassicSimilarity], result of:
              0.2101263 = score(doc=306,freq=2.0), product of:
                0.32046703 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.037799787 = queryNorm
                0.65568775 = fieldWeight in 306, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=306)
          0.33333334 = coord(1/3)
        0.2101263 = weight(_text_:2f in 306) [ClassicSimilarity], result of:
          0.2101263 = score(doc=306,freq=2.0), product of:
            0.32046703 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.037799787 = queryNorm
            0.65568775 = fieldWeight in 306, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0546875 = fieldNorm(doc=306)
      0.4 = coord(2/5)
    
    Content
    Vgl.: http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=5386707&url=http%3A%2F%2Fieeexplore.ieee.org%2Fxpls%2Fabs_all.jsp%3Farnumber%3D5386707.
  5. Donsbach, W.: Wahrheit in den Medien : über den Sinn eines methodischen Objektivitätsbegriffes (2001) 0.08
    0.08004812 = product of:
      0.2001203 = sum of:
        0.050030075 = product of:
          0.15009022 = sum of:
            0.15009022 = weight(_text_:3a in 5895) [ClassicSimilarity], result of:
              0.15009022 = score(doc=5895,freq=2.0), product of:
                0.32046703 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.037799787 = queryNorm
                0.46834838 = fieldWeight in 5895, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5895)
          0.33333334 = coord(1/3)
        0.15009022 = weight(_text_:2f in 5895) [ClassicSimilarity], result of:
          0.15009022 = score(doc=5895,freq=2.0), product of:
            0.32046703 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.037799787 = queryNorm
            0.46834838 = fieldWeight in 5895, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5895)
      0.4 = coord(2/5)
    
    Source
    Politische Meinung. 381(2001) Nr.1, S.65-74 [https%3A%2F%2Fwww.dgfe.de%2Ffileadmin%2FOrdnerRedakteure%2FSektionen%2FSek02_AEW%2FKWF%2FPublikationen_Reihe_1989-2003%2FBand_17%2FBd_17_1994_355-406_A.pdf&usg=AOvVaw2KcbRsHy5UQ9QRIUyuOLNi]
  6. Mudge, S.; Hoek, D.J.: Describing jazz, blues, and popular 78 rpm sound recordings : suggestions and guidelines (2000) 0.08
    0.07935123 = product of:
      0.19837807 = sum of:
        0.18631978 = weight(_text_:discs in 965) [ClassicSimilarity], result of:
          0.18631978 = score(doc=965,freq=2.0), product of:
            0.30176762 = queryWeight, product of:
              7.983315 = idf(docFreq=40, maxDocs=44218)
              0.037799787 = queryNorm
            0.617428 = fieldWeight in 965, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.983315 = idf(docFreq=40, maxDocs=44218)
              0.0546875 = fieldNorm(doc=965)
        0.012058287 = product of:
          0.03617486 = sum of:
            0.03617486 = weight(_text_:29 in 965) [ClassicSimilarity], result of:
              0.03617486 = score(doc=965,freq=2.0), product of:
                0.13296783 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037799787 = queryNorm
                0.27205724 = fieldWeight in 965, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=965)
          0.33333334 = coord(1/3)
      0.4 = coord(2/5)
    
    Abstract
    Since 78 rpm sound recordings of jazz, blues, and popular music are today a specialized medium, they receive limited attention in cataloging rules and guides. The cataloging of 78 rpm discs at Indiana University's Archives of Traditional Music is based on established standards; nevertheless, certain local decisions are necessary when general rules are not clear. The Archives' decisions related to the description of their 78 rpm collections are explained and presented with examples in MARC format, and issues of access related to the choice of main entry are also covered
    Source
    Cataloging and classification quarterly. 29(2000) no.3, S.21-47
  7. San Segundo, R.: ¬A new conception of representation of knowledge (2004) 0.04
    0.041020717 = product of:
      0.10255179 = sum of:
        0.09572332 = weight(_text_:compact in 3077) [ClassicSimilarity], result of:
          0.09572332 = score(doc=3077,freq=2.0), product of:
            0.28613505 = queryWeight, product of:
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.037799787 = queryNorm
            0.33453897 = fieldWeight in 3077, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.03125 = fieldNorm(doc=3077)
        0.006828466 = product of:
          0.020485397 = sum of:
            0.020485397 = weight(_text_:22 in 3077) [ClassicSimilarity], result of:
              0.020485397 = score(doc=3077,freq=2.0), product of:
                0.13236842 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037799787 = queryNorm
                0.15476047 = fieldWeight in 3077, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3077)
          0.33333334 = coord(1/3)
      0.4 = coord(2/5)
    
    Abstract
    The new term Representation of knowledge, applied to the framework of electronic segments of information, with comprehension of new material support for information, and a review and total conceptualisation of the terminology which is being applied, entails a review of all traditional documentary practices. Therefore, a definition of the concept of Representation of knowledge is indispensable. The term representation has been used in westere cultural and intellectual tradition to refer to the diverse ways that a subject comprehends an object. Representation is a process which requires the structure of natural language and human memory whereby it is interwoven in a subject and in conscience. However, at the present time, the term Representation of knowledge is applied to the processing of electronic information, combined with the aim of emulating the human mind in such a way that one has endeavoured to transfer, with great difficulty, the complex structurality of the conceptual representation of human knowledge to new digital information technologies. Thus, nowadays, representation of knowledge has taken an diverse meanings and it has focussed, for the moment, an certain structures and conceptual hierarchies which carry and transfer information, and has initially been based an the current representation of knowledge using artificial intelligence. The traditional languages of documentation, also referred to as languages of representation, offer a structured representation of conceptual fields, symbols and terms of natural and notational language, and they are the pillars for the necessary correspondence between the object or text and its representation. These correspondences, connections and symbolisations will be established within the electronic framework by means of different models and of the "goal" domain, which will give rise to organisations, structures, maps, networks and levels, as new electronic documents are not compact units but segments of information. Thus, the new representation of knowledge refers to data, images, figures and symbolised, treated, processed and structured ideas which replace or refer to documents within the framework of technical processing and the recuperation of electronic information.
    Date
    2. 1.2005 18:22:25
  8. Horch, H.O.: Compact Memory : Ein DFG-Projekt zur retrospektiven Digitalisierung jüdischer Periodika im deutschsprachigen Raum (2006) 0.03
    0.033503164 = product of:
      0.16751581 = sum of:
        0.16751581 = weight(_text_:compact in 49) [ClassicSimilarity], result of:
          0.16751581 = score(doc=49,freq=2.0), product of:
            0.28613505 = queryWeight, product of:
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.037799787 = queryNorm
            0.5854432 = fieldWeight in 49, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.0546875 = fieldNorm(doc=49)
      0.2 = coord(1/5)
    
  9. Soricut, R.; Marcu, D.: Abstractive headline generation using WIDL-expressions (2007) 0.03
    0.033503164 = product of:
      0.16751581 = sum of:
        0.16751581 = weight(_text_:compact in 943) [ClassicSimilarity], result of:
          0.16751581 = score(doc=943,freq=2.0), product of:
            0.28613505 = queryWeight, product of:
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.037799787 = queryNorm
            0.5854432 = fieldWeight in 943, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.0546875 = fieldNorm(doc=943)
      0.2 = coord(1/5)
    
    Abstract
    We present a new paradigm for the automatic creation of document headlines that is based on direct transformation of relevant textual information into well-formed textual output. Starting from an input document, we automatically create compact representations of weighted finite sets of strings, called WIDL-expressions, which encode the most important topics in the document. A generic natural language generation engine performs the headline generation task, driven by both statistical knowledge encapsulated in WIDL-expressions (representing topic biases induced by the input document) and statistical knowledge encapsulated in language models (representing biases induced by the target language). Our evaluation shows similar performance in quality with a state-of-the-art, extractive approach to headline generation, and significant improvements in quality over previously proposed solutions to abstractive headline generation.
  10. Roosa, M.: Sound and audio archives (2009) 0.03
    0.031940535 = product of:
      0.15970267 = sum of:
        0.15970267 = weight(_text_:discs in 3883) [ClassicSimilarity], result of:
          0.15970267 = score(doc=3883,freq=2.0), product of:
            0.30176762 = queryWeight, product of:
              7.983315 = idf(docFreq=40, maxDocs=44218)
              0.037799787 = queryNorm
            0.52922404 = fieldWeight in 3883, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.983315 = idf(docFreq=40, maxDocs=44218)
              0.046875 = fieldNorm(doc=3883)
      0.2 = coord(1/5)
    
    Abstract
    This entry provides an overview of sound archives, including reasons why sound archives exist, their history and organization; types of sound recordings collected, methods of description, and access to and preservation of recorded sound materials. Recorded sound formats (e.g., cylinders, discs, long playing (LP) records, etc.) are covered in the context of how they have been (and are currently being) collected, described, preserved, and made available. Select projects and programs undertaken by regional, special, and national sound archives are covered. Professional associations that focus on sound archiving are described as are funding avenues for sound archives. Description is also included of work being carried out in the United States to modify copyright law to better enable sound archives to preserve their holdings for future generations of users.
  11. Egghe, L.; Rousseau, R.: ¬A measure for the cohesion of weighted networks (2003) 0.02
    0.02393083 = product of:
      0.11965415 = sum of:
        0.11965415 = weight(_text_:compact in 5157) [ClassicSimilarity], result of:
          0.11965415 = score(doc=5157,freq=2.0), product of:
            0.28613505 = queryWeight, product of:
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.037799787 = queryNorm
            0.4181737 = fieldWeight in 5157, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5157)
      0.2 = coord(1/5)
    
    Abstract
    Measurement of the degree of interconnectedness in graph like networks of hyperlinks or citations can indicate the existence of research fields and assist in comparative evaluation of research efforts. In this issue we begin with Egghe and Rousseau who review compactness measures and investigate the compactness of a network as a weighted graph with dissimilarity values characterizing the arcs between nodes. They make use of a generalization of the Botofogo, Rivlin, Shneiderman, (BRS) compaction measure which treats the distance between unreachable nodes not as infinity but rather as the number of nodes in the network. The dissimilarity values are determined by summing the reciprocals of the weights of the arcs in the shortest chain between two nodes where no weight is smaller than one. The BRS measure is then the maximum value for the sum of the dissimilarity measures less the actual sum divided by the difference between the maximum and minimum. The Wiener index, the sum of all elements in the dissimilarity matrix divided by two, is then computed for Small's particle physics co-citation data as well as the BRS measure, the dissimilarity values and shortest paths. The compactness measure for the weighted network is smaller than for the un-weighted. When the bibliographic coupling network is utilized it is shown to be less compact than the co-citation network which indicates that the new measure produces results that confirm to an obvious case.
  12. Na, S.-H.; Kang, I.-S.; Lee, J.-H.: Parsimonious translation models for information retrieval (2007) 0.02
    0.02393083 = product of:
      0.11965415 = sum of:
        0.11965415 = weight(_text_:compact in 898) [ClassicSimilarity], result of:
          0.11965415 = score(doc=898,freq=2.0), product of:
            0.28613505 = queryWeight, product of:
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.037799787 = queryNorm
            0.4181737 = fieldWeight in 898, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.0390625 = fieldNorm(doc=898)
      0.2 = coord(1/5)
    
    Abstract
    In the KL divergence framework, the extended language modeling approach has a critical problem of estimating a query model, which is the probabilistic model that encodes the user's information need. For query expansion in initial retrieval, the translation model had been proposed to involve term co-occurrence statistics. However, the translation model was difficult to apply, because the term co-occurrence statistics must be constructed in the offline time. Especially in a large collection, constructing such a large matrix of term co-occurrences statistics prohibitively increases time and space complexity. In addition, reliable retrieval performance cannot be guaranteed because the translation model may comprise noisy non-topical terms in documents. To resolve these problems, this paper investigates an effective method to construct co-occurrence statistics and eliminate noisy terms by employing a parsimonious translation model. The parsimonious translation model is a compact version of a translation model that can reduce the number of terms containing non-zero probabilities by eliminating non-topical terms in documents. Through experimentation on seven different test collections, we show that the query model estimated from the parsimonious translation model significantly outperforms not only the baseline language modeling, but also the non-parsimonious models.
  13. Hoeber, O.; Yang, X.D.: HotMap : supporting visual exploration of Web search results (2009) 0.02
    0.02393083 = product of:
      0.11965415 = sum of:
        0.11965415 = weight(_text_:compact in 2700) [ClassicSimilarity], result of:
          0.11965415 = score(doc=2700,freq=2.0), product of:
            0.28613505 = queryWeight, product of:
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.037799787 = queryNorm
            0.4181737 = fieldWeight in 2700, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2700)
      0.2 = coord(1/5)
    
    Abstract
    Although information retrieval techniques used by Web search engines have improved substantially over the years, the results of Web searches have continued to be represented in simple list-based formats. Although the list-based representation makes it easy to evaluate a single document for relevance, it does not support the users in the broader tasks of manipulating or exploring the search results as they attempt to find a collection of relevant documents. HotMap is a meta-search system that provides a compact visual representation of Web search results at two levels of detail, and it supports interactive exploration via nested sorting of Web search results based on query term frequencies. An evaluation of the search results for a set of vague queries has shown that the re-sorted search results can provide a higher portion of relevant documents among the top search results. User studies show an increase in speed and effectiveness and a reduction in missed documents when comparing HotMap to the list-based representation used by Google. Subjective measures were positive, and users showed a preference for the HotMap interface. These results provide evidence for the utility of next-generation Web search results interfaces that promote interactive search results exploration.
  14. Tochtermann, K.; Granitzer, M.: Wissenserschließung : Pfade durch den digitalen Informationsdschungel (2005) 0.02
    0.021950265 = product of:
      0.10975133 = sum of:
        0.10975133 = product of:
          0.16462699 = sum of:
            0.082685396 = weight(_text_:29 in 4390) [ClassicSimilarity], result of:
              0.082685396 = score(doc=4390,freq=2.0), product of:
                0.13296783 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037799787 = queryNorm
                0.6218451 = fieldWeight in 4390, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.125 = fieldNorm(doc=4390)
            0.08194159 = weight(_text_:22 in 4390) [ClassicSimilarity], result of:
              0.08194159 = score(doc=4390,freq=2.0), product of:
                0.13236842 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037799787 = queryNorm
                0.61904186 = fieldWeight in 4390, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=4390)
          0.6666667 = coord(2/3)
      0.2 = coord(1/5)
    
    Date
    5.11.2005 19:18:22
    Source
    Wissensmanagement. 7(2005) H.5, S.26-29
  15. Jörn, F.: Wie Google für uns nach der ominösen Gluonenkraft stöbert : Software-Krabbler machen sich vor der Anfrage auf die Suche - Das Netz ist etwa fünfhundertmal größer als alles Durchforschte (2001) 0.02
    0.020510359 = product of:
      0.051275894 = sum of:
        0.04786166 = weight(_text_:compact in 3684) [ClassicSimilarity], result of:
          0.04786166 = score(doc=3684,freq=2.0), product of:
            0.28613505 = queryWeight, product of:
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.037799787 = queryNorm
            0.16726948 = fieldWeight in 3684, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.015625 = fieldNorm(doc=3684)
        0.003414233 = product of:
          0.010242699 = sum of:
            0.010242699 = weight(_text_:22 in 3684) [ClassicSimilarity], result of:
              0.010242699 = score(doc=3684,freq=2.0), product of:
                0.13236842 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037799787 = queryNorm
                0.07738023 = fieldWeight in 3684, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.015625 = fieldNorm(doc=3684)
          0.33333334 = coord(1/3)
      0.4 = coord(2/5)
    
    Abstract
    Es gibt riesige Inseln nach außen nicht verlinkter Dateien, es gibt Formate, die dem HTML-Standard nicht entsprechen und von Suchmaschinen nicht oder ungern gelesen werden, von Word-Dokumenten bis zu PDF-Dateien (Google durchkämmt sie und speichert sie zum schnellen Uberblick und vorteilhaft kopierbar als Textdateien!), Tabellen und Folienvorträge, Gedcom-Stammbäume, vor allem aber Bilder, Filme, Musik, die sich nur schwer elektronisch katalogisieren lassen. Haben Suchmaschinen Zeit, mit künstlicher Intelligenz herauszufinden, ob auf einem Bild eine Person ist? Und wenn, wer mag es sein? Infoseek bemüht sich in einer eigenen Bildersuche darum, kann allerdings auch kein Konterfei von Luis Trenker oder Toni Sailer herbeizaubern, wogegen "Luis Trenker Bild", besonders bei Google, zum Foto führt. "Britney Spears" strahlt einem gleich entgegen! Wenn Bilder beliebig benannt werden, bleiben sie unauffindbar. MP3-Dateien enthalten oft maschinenlesbar den Titel in der Musikdatei - eine große Hilfe für Suchmaschinen. Neue Webformate wie Macromedia Flash, dem Internet-Veteranen ohnehin ein Graus, vernebeln das in ihrem Troß Folgende. Und bietet eine Internetseite eine eigene Datenbanksuche an, dann bleibt diese Datenbank vor Suchmaschinen verborgen, von Telefonnummern und Zügen bis zu Artikeln dieser Zeitung. Zuvorkommender ist es, die Inhalte in Hypertext ins Netz zu stellen - für die Suchmaschinen und zusätzlich manuell darin suchen zu lassen. Suchmaschinen wie Freefind oder Atomz bieten das kostenlos an. Grundsätzlich können Suchmaschinen kostenpflichtige Inhalte nicht durchkämmen. So wie sich die olympische Idee inzwischen den Profis gebeugt hat, besteht auch im Internet die Gefahr, daß es immer kommerzieller zugeht. Ein Musterbeispiel sind WapInhalte für mobile Betrachter, die im Gegensatz zu HTML-Seiten nicht systematisch von einem Domänennamen her über Links erreichbar sind. Wap-Suchmaschinen weisen also nur angemeldete Seiten nach und spielen eine untergeordnete Rolle. Viel lieber schleusen die Mobilfunkanbieter ihre Kunden über Portale. Zollund Zahlgrenzen, Partikularismus zerstören das Netz. Beim japanischen Imode, mit HTML kompatibel, ist das anders; selbst Google bietet über www.google.com/imode Suche an, hat dann aber Mühe, Imode-Inhalte (in cHTML, compact HTML) von HTML zu unterscheiden. Grundsätzlich ist die Rivalität zwischen Internet-Portalen mit ihrer Zugangsführung und Suchmaschinen für Quereinsteiger noch nicht ausgefochten. Noch aus der Vor-Web-Zeit stammen Diskussionsforen. Dort werden zu bestimmten Themen Meinungen ausgetauscht, - moderiert oder wildwachsend.
    Date
    22. 6.2005 9:52:00
  16. Herrmann, C.: Partikulare Konkretion universal zugänglicher Information : Beobachtungen zur Konzeptionierung fachlicher Internet-Seiten am Beispiel der Theologie (2000) 0.02
    0.019206481 = product of:
      0.0960324 = sum of:
        0.0960324 = product of:
          0.1440486 = sum of:
            0.07234972 = weight(_text_:29 in 4364) [ClassicSimilarity], result of:
              0.07234972 = score(doc=4364,freq=2.0), product of:
                0.13296783 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037799787 = queryNorm
                0.5441145 = fieldWeight in 4364, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4364)
            0.07169889 = weight(_text_:22 in 4364) [ClassicSimilarity], result of:
              0.07169889 = score(doc=4364,freq=2.0), product of:
                0.13236842 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037799787 = queryNorm
                0.5416616 = fieldWeight in 4364, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4364)
          0.6666667 = coord(2/3)
      0.2 = coord(1/5)
    
    Date
    22. 1.2000 19:29:08
  17. Weinberg, B.H.: Book indexes in France : medieval specimens and modern practices (2000) 0.02
    0.019206481 = product of:
      0.0960324 = sum of:
        0.0960324 = product of:
          0.1440486 = sum of:
            0.07234972 = weight(_text_:29 in 486) [ClassicSimilarity], result of:
              0.07234972 = score(doc=486,freq=2.0), product of:
                0.13296783 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037799787 = queryNorm
                0.5441145 = fieldWeight in 486, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.109375 = fieldNorm(doc=486)
            0.07169889 = weight(_text_:22 in 486) [ClassicSimilarity], result of:
              0.07169889 = score(doc=486,freq=2.0), product of:
                0.13236842 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037799787 = queryNorm
                0.5416616 = fieldWeight in 486, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=486)
          0.6666667 = coord(2/3)
      0.2 = coord(1/5)
    
    Date
    20. 4.2002 19:29:54
    Source
    Indexer. 22(2000) no.1, S.2-13
  18. Mauer, P.: Embedded indexing : pros and cons for the indexer (2000) 0.02
    0.019206481 = product of:
      0.0960324 = sum of:
        0.0960324 = product of:
          0.1440486 = sum of:
            0.07234972 = weight(_text_:29 in 488) [ClassicSimilarity], result of:
              0.07234972 = score(doc=488,freq=2.0), product of:
                0.13296783 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.037799787 = queryNorm
                0.5441145 = fieldWeight in 488, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.109375 = fieldNorm(doc=488)
            0.07169889 = weight(_text_:22 in 488) [ClassicSimilarity], result of:
              0.07169889 = score(doc=488,freq=2.0), product of:
                0.13236842 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.037799787 = queryNorm
                0.5416616 = fieldWeight in 488, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=488)
          0.6666667 = coord(2/3)
      0.2 = coord(1/5)
    
    Date
    21. 4.2002 9:29:38
    Source
    Indexer. 22(2000) no.1, S.27-28
  19. Dominich, S.; Kiezer, T.: ¬A measure theoretic approach to information retrieval (2007) 0.02
    0.019144665 = product of:
      0.09572332 = sum of:
        0.09572332 = weight(_text_:compact in 445) [ClassicSimilarity], result of:
          0.09572332 = score(doc=445,freq=2.0), product of:
            0.28613505 = queryWeight, product of:
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.037799787 = queryNorm
            0.33453897 = fieldWeight in 445, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.03125 = fieldNorm(doc=445)
      0.2 = coord(1/5)
    
    Abstract
    The vector space model of information retrieval is one of the classical and widely applied retrieval models. Paradoxically, it has been characterized by a discrepancy between its formal framework and implementable form. The underlying concepts of the vector space model are mathematical terms: linear space, vector, and inner product. However, in the vector space model, the mathematical meaning of these concepts is not preserved. They are used as mere computational constructs or metaphors. Thus, the vector space model actually does not follow formally from the mathematical concepts on which it has been claimed to rest. This problem has been recognized for more than two decades, but no proper solution has emerged so far. The present article proposes a solution to this problem. First, the concept of retrieval is defined based on the mathematical measure theory. Then, retrieval is particularized using fuzzy set theory. As a result, the retrieval function is conceived as the cardinality of the intersection of two fuzzy sets. This view makes it possible to build a connection to linear spaces. It is shown that the classical and the generalized vector space models, as well as the latent semantic indexing model, gain a correct formal background with which they are consistent. At the same time it becomes clear that the inner product is not a necessary ingredient of the vector space model, and hence of Information Retrieval (IR). The Principle of Object Invariance is introduced to handle this situation. Moreover, this view makes it possible to consistently formulate new retrieval methods: in linear space with general basis, entropy-based, and probability-based. It is also shown that Information Retrieval may be viewed as integral calculus, and thus it gains a very compact and elegant mathematical way of writing. Also, Information Retrieval may thus be conceived as an application of mathematical measure theory.
  20. Dunlavy, D.M.; O'Leary, D.P.; Conroy, J.M.; Schlesinger, J.D.: QCS: A system for querying, clustering and summarizing documents (2007) 0.02
    0.019144665 = product of:
      0.09572332 = sum of:
        0.09572332 = weight(_text_:compact in 947) [ClassicSimilarity], result of:
          0.09572332 = score(doc=947,freq=2.0), product of:
            0.28613505 = queryWeight, product of:
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.037799787 = queryNorm
            0.33453897 = fieldWeight in 947, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.5697527 = idf(docFreq=61, maxDocs=44218)
              0.03125 = fieldNorm(doc=947)
      0.2 = coord(1/5)
    
    Abstract
    Information retrieval systems consist of many complicated components. Research and development of such systems is often hampered by the difficulty in evaluating how each particular component would behave across multiple systems. We present a novel integrated information retrieval system-the Query, Cluster, Summarize (QCS) system-which is portable, modular, and permits experimentation with different instantiations of each of the constituent text analysis components. Most importantly, the combination of the three types of methods in the QCS design improves retrievals by providing users more focused information organized by topic. We demonstrate the improved performance by a series of experiments using standard test sets from the Document Understanding Conferences (DUC) as measured by the best known automatic metric for summarization system evaluation, ROUGE. Although the DUC data and evaluations were originally designed to test multidocument summarization, we developed a framework to extend it to the task of evaluation for each of the three components: query, clustering, and summarization. Under this framework, we then demonstrate that the QCS system (end-to-end) achieves performance as good as or better than the best summarization engines. Given a query, QCS retrieves relevant documents, separates the retrieved documents into topic clusters, and creates a single summary for each cluster. In the current implementation, Latent Semantic Indexing is used for retrieval, generalized spherical k-means is used for the document clustering, and a method coupling sentence "trimming" and a hidden Markov model, followed by a pivoted QR decomposition, is used to create a single extract summary for each cluster. The user interface is designed to provide access to detailed information in a compact and useful format. Our system demonstrates the feasibility of assembling an effective IR system from existing software libraries, the usefulness of the modularity of the design, and the value of this particular combination of modules.

Languages

Types

Themes