Search (7249 results, page 1 of 363)

  • × year_i:[2000 TO 2010}
  1. Schrodt, R.: Tiefen und Untiefen im wissenschaftlichen Sprachgebrauch (2008) 0.36
    0.3646155 = product of:
      0.82038486 = sum of:
        0.082038485 = product of:
          0.24611545 = sum of:
            0.24611545 = weight(_text_:3a in 140) [ClassicSimilarity], result of:
              0.24611545 = score(doc=140,freq=2.0), product of:
                0.32843533 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.038739666 = queryNorm
                0.7493574 = fieldWeight in 140, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0625 = fieldNorm(doc=140)
          0.33333334 = coord(1/3)
        0.24611545 = weight(_text_:2f in 140) [ClassicSimilarity], result of:
          0.24611545 = score(doc=140,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.7493574 = fieldWeight in 140, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0625 = fieldNorm(doc=140)
        0.24611545 = weight(_text_:2f in 140) [ClassicSimilarity], result of:
          0.24611545 = score(doc=140,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.7493574 = fieldWeight in 140, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0625 = fieldNorm(doc=140)
        0.24611545 = weight(_text_:2f in 140) [ClassicSimilarity], result of:
          0.24611545 = score(doc=140,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.7493574 = fieldWeight in 140, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0625 = fieldNorm(doc=140)
      0.44444445 = coord(4/9)
    
    Content
    Vgl. auch: https://studylibde.com/doc/13053640/richard-schrodt. Vgl. auch: http%3A%2F%2Fwww.univie.ac.at%2FGermanistik%2Fschrodt%2Fvorlesung%2Fwissenschaftssprache.doc&usg=AOvVaw1lDLDR6NFf1W0-oC9mEUJf.
  2. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.35
    0.35057482 = product of:
      0.6310347 = sum of:
        0.06152886 = product of:
          0.18458658 = sum of:
            0.18458658 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
              0.18458658 = score(doc=562,freq=2.0), product of:
                0.32843533 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.038739666 = queryNorm
                0.56201804 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
        0.18458658 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.18458658 = score(doc=562,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.18458658 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.18458658 = score(doc=562,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.18458658 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.18458658 = score(doc=562,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.01574607 = product of:
          0.03149214 = sum of:
            0.03149214 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
              0.03149214 = score(doc=562,freq=2.0), product of:
                0.13565971 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038739666 = queryNorm
                0.23214069 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.5 = coord(1/2)
      0.5555556 = coord(5/9)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  3. Mas, S.; Marleau, Y.: Proposition of a faceted classification model to support corporate information organization and digital records management (2009) 0.35
    0.34804496 = product of:
      0.6264809 = sum of:
        0.06152886 = product of:
          0.18458658 = sum of:
            0.18458658 = weight(_text_:3a in 2918) [ClassicSimilarity], result of:
              0.18458658 = score(doc=2918,freq=2.0), product of:
                0.32843533 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.038739666 = queryNorm
                0.56201804 = fieldWeight in 2918, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2918)
          0.33333334 = coord(1/3)
        0.18458658 = weight(_text_:2f in 2918) [ClassicSimilarity], result of:
          0.18458658 = score(doc=2918,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.56201804 = fieldWeight in 2918, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=2918)
        0.011192262 = weight(_text_:information in 2918) [ClassicSimilarity], result of:
          0.011192262 = score(doc=2918,freq=4.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.16457605 = fieldWeight in 2918, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2918)
        0.18458658 = weight(_text_:2f in 2918) [ClassicSimilarity], result of:
          0.18458658 = score(doc=2918,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.56201804 = fieldWeight in 2918, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=2918)
        0.18458658 = weight(_text_:2f in 2918) [ClassicSimilarity], result of:
          0.18458658 = score(doc=2918,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.56201804 = fieldWeight in 2918, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=2918)
      0.5555556 = coord(5/9)
    
    Abstract
    The employees of an organization often use a personal hierarchical classification scheme to organize digital documents that are stored on their own workstations. As this may make it hard for other employees to retrieve these documents, there is a risk that the organization will lose track of needed documentation. Furthermore, the inherent boundaries of such a hierarchical structure require making arbitrary decisions about which specific criteria the classification will b.e based on (for instance, the administrative activity or the document type, although a document can have several attributes and require classification in several classes).A faceted classification model to support corporate information organization is proposed. Partially based on Ranganathan's facets theory, this model aims not only to standardize the organization of digital documents, but also to simplify the management of a document throughout its life cycle for both individuals and organizations, while ensuring compliance to regulatory and policy requirements.
    Footnote
    Vgl.: http://ieeexplore.ieee.org/Xplore/login.jsp?reload=true&url=http%3A%2F%2Fieeexplore.ieee.org%2Fiel5%2F4755313%2F4755314%2F04755480.pdf%3Farnumber%3D4755480&authDecision=-203.
  4. Stojanovic, N.: Ontology-based Information Retrieval : methods and tools for cooperative query answering (2005) 0.32
    0.32309067 = product of:
      0.484636 = sum of:
        0.041019242 = product of:
          0.12305772 = sum of:
            0.12305772 = weight(_text_:3a in 701) [ClassicSimilarity], result of:
              0.12305772 = score(doc=701,freq=2.0), product of:
                0.32843533 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.038739666 = queryNorm
                0.3746787 = fieldWeight in 701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
          0.33333334 = coord(1/3)
        0.12305772 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.12305772 = score(doc=701,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.015828248 = weight(_text_:information in 701) [ClassicSimilarity], result of:
          0.015828248 = score(doc=701,freq=18.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.23274568 = fieldWeight in 701, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.058615338 = weight(_text_:retrieval in 701) [ClassicSimilarity], result of:
          0.058615338 = score(doc=701,freq=28.0), product of:
            0.1171842 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.038739666 = queryNorm
            0.5001983 = fieldWeight in 701, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.12305772 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.12305772 = score(doc=701,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.12305772 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.12305772 = score(doc=701,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
      0.6666667 = coord(6/9)
    
    Abstract
    By the explosion of possibilities for a ubiquitous content production, the information overload problem reaches the level of complexity which cannot be managed by traditional modelling approaches anymore. Due to their pure syntactical nature traditional information retrieval approaches did not succeed in treating content itself (i.e. its meaning, and not its representation). This leads to a very low usefulness of the results of a retrieval process for a user's task at hand. In the last ten years ontologies have been emerged from an interesting conceptualisation paradigm to a very promising (semantic) modelling technology, especially in the context of the Semantic Web. From the information retrieval point of view, ontologies enable a machine-understandable form of content description, such that the retrieval process can be driven by the meaning of the content. However, the very ambiguous nature of the retrieval process in which a user, due to the unfamiliarity with the underlying repository and/or query syntax, just approximates his information need in a query, implies a necessity to include the user in the retrieval process more actively in order to close the gap between the meaning of the content and the meaning of a user's query (i.e. his information need). This thesis lays foundation for such an ontology-based interactive retrieval process, in which the retrieval system interacts with a user in order to conceptually interpret the meaning of his query, whereas the underlying domain ontology drives the conceptualisation process. In that way the retrieval process evolves from a query evaluation process into a highly interactive cooperation between a user and the retrieval system, in which the system tries to anticipate the user's information need and to deliver the relevant content proactively. Moreover, the notion of content relevance for a user's query evolves from a content dependent artefact to the multidimensional context-dependent structure, strongly influenced by the user's preferences. This cooperation process is realized as the so-called Librarian Agent Query Refinement Process. In order to clarify the impact of an ontology on the retrieval process (regarding its complexity and quality), a set of methods and tools for different levels of content and query formalisation is developed, ranging from pure ontology-based inferencing to keyword-based querying in which semantics automatically emerges from the results. Our evaluation studies have shown that the possibilities to conceptualize a user's information need in the right manner and to interpret the retrieval results accordingly are key issues for realizing much more meaningful information retrieval systems.
    Content
    Vgl.: http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F1627&ei=tAtYUYrBNoHKtQb3l4GYBw&usg=AFQjCNHeaxKkKU3-u54LWxMNYGXaaDLCGw&sig2=8WykXWQoDKjDSdGtAakH2Q&bvm=bv.44442042,d.Yms.
  5. Vetere, G.; Lenzerini, M.: Models for semantic interoperability in service-oriented architectures (2005) 0.32
    0.31903857 = product of:
      0.7178368 = sum of:
        0.07178368 = product of:
          0.21535103 = sum of:
            0.21535103 = weight(_text_:3a in 306) [ClassicSimilarity], result of:
              0.21535103 = score(doc=306,freq=2.0), product of:
                0.32843533 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.038739666 = queryNorm
                0.65568775 = fieldWeight in 306, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=306)
          0.33333334 = coord(1/3)
        0.21535103 = weight(_text_:2f in 306) [ClassicSimilarity], result of:
          0.21535103 = score(doc=306,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.65568775 = fieldWeight in 306, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0546875 = fieldNorm(doc=306)
        0.21535103 = weight(_text_:2f in 306) [ClassicSimilarity], result of:
          0.21535103 = score(doc=306,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.65568775 = fieldWeight in 306, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0546875 = fieldNorm(doc=306)
        0.21535103 = weight(_text_:2f in 306) [ClassicSimilarity], result of:
          0.21535103 = score(doc=306,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.65568775 = fieldWeight in 306, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0546875 = fieldNorm(doc=306)
      0.44444445 = coord(4/9)
    
    Content
    Vgl.: http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=5386707&url=http%3A%2F%2Fieeexplore.ieee.org%2Fxpls%2Fabs_all.jsp%3Farnumber%3D5386707.
  6. Donsbach, W.: Wahrheit in den Medien : über den Sinn eines methodischen Objektivitätsbegriffes (2001) 0.29
    0.2885198 = product of:
      0.5193356 = sum of:
        0.051274054 = product of:
          0.15382215 = sum of:
            0.15382215 = weight(_text_:3a in 5895) [ClassicSimilarity], result of:
              0.15382215 = score(doc=5895,freq=2.0), product of:
                0.32843533 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.038739666 = queryNorm
                0.46834838 = fieldWeight in 5895, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5895)
          0.33333334 = coord(1/3)
        0.15382215 = weight(_text_:2f in 5895) [ClassicSimilarity], result of:
          0.15382215 = score(doc=5895,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.46834838 = fieldWeight in 5895, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5895)
        0.0065951035 = weight(_text_:information in 5895) [ClassicSimilarity], result of:
          0.0065951035 = score(doc=5895,freq=2.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.09697737 = fieldWeight in 5895, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5895)
        0.15382215 = weight(_text_:2f in 5895) [ClassicSimilarity], result of:
          0.15382215 = score(doc=5895,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.46834838 = fieldWeight in 5895, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5895)
        0.15382215 = weight(_text_:2f in 5895) [ClassicSimilarity], result of:
          0.15382215 = score(doc=5895,freq=2.0), product of:
            0.32843533 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.038739666 = queryNorm
            0.46834838 = fieldWeight in 5895, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5895)
      0.5555556 = coord(5/9)
    
    Source
    Politische Meinung. 381(2001) Nr.1, S.65-74 [https%3A%2F%2Fwww.dgfe.de%2Ffileadmin%2FOrdnerRedakteure%2FSektionen%2FSek02_AEW%2FKWF%2FPublikationen_Reihe_1989-2003%2FBand_17%2FBd_17_1994_355-406_A.pdf&usg=AOvVaw2KcbRsHy5UQ9QRIUyuOLNi]
    Theme
    Information
  7. Buxton, A.; Hopkinson, A.: ¬The CDS/ISIS for Windows handbook (2001) 0.11
    0.10619919 = product of:
      0.31859756 = sum of:
        0.1903564 = weight(_text_:line in 775) [ClassicSimilarity], result of:
          0.1903564 = score(doc=775,freq=4.0), product of:
            0.21724595 = queryWeight, product of:
              5.6078424 = idf(docFreq=440, maxDocs=44218)
              0.038739666 = queryNorm
            0.87622535 = fieldWeight in 775, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.6078424 = idf(docFreq=440, maxDocs=44218)
              0.078125 = fieldNorm(doc=775)
        0.03230928 = weight(_text_:information in 775) [ClassicSimilarity], result of:
          0.03230928 = score(doc=775,freq=12.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.47509015 = fieldWeight in 775, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=775)
        0.095931865 = weight(_text_:retrieval in 775) [ClassicSimilarity], result of:
          0.095931865 = score(doc=775,freq=12.0), product of:
            0.1171842 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.038739666 = queryNorm
            0.81864166 = fieldWeight in 775, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.078125 = fieldNorm(doc=775)
      0.33333334 = coord(3/9)
    
    COMPASS
    Information retrieval / Use of / On-line computers
    LCSH
    ISIS (Information retrieval system) / Handbooks, manuals, etc.
    Information storage and retrieval systems / Handbooks, manuals, etc.
    Subject
    ISIS (Information retrieval system) / Handbooks, manuals, etc.
    Information storage and retrieval systems / Handbooks, manuals, etc.
    Information retrieval / Use of / On-line computers
  8. Greiff, W.R.: ¬The use of exploratory data analysis in information retrieval research (2000) 0.09
    0.09224124 = product of:
      0.20754279 = sum of:
        0.08076138 = weight(_text_:line in 32) [ClassicSimilarity], result of:
          0.08076138 = score(doc=32,freq=2.0), product of:
            0.21724595 = queryWeight, product of:
              5.6078424 = idf(docFreq=440, maxDocs=44218)
              0.038739666 = queryNorm
            0.37175092 = fieldWeight in 32, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.6078424 = idf(docFreq=440, maxDocs=44218)
              0.046875 = fieldNorm(doc=32)
        0.019385567 = weight(_text_:information in 32) [ClassicSimilarity], result of:
          0.019385567 = score(doc=32,freq=12.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.2850541 = fieldWeight in 32, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=32)
        0.057559118 = weight(_text_:retrieval in 32) [ClassicSimilarity], result of:
          0.057559118 = score(doc=32,freq=12.0), product of:
            0.1171842 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.038739666 = queryNorm
            0.49118498 = fieldWeight in 32, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=32)
        0.049836725 = weight(_text_:techniques in 32) [ClassicSimilarity], result of:
          0.049836725 = score(doc=32,freq=2.0), product of:
            0.17065717 = queryWeight, product of:
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.038739666 = queryNorm
            0.2920283 = fieldWeight in 32, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.046875 = fieldNorm(doc=32)
      0.44444445 = coord(4/9)
    
    Abstract
    We report on a line of work in which techniques of Exploratory Data Analysis (EDA) have been used as a vehicle for better understanding of the issues confronting the researcher in information retrieval (IR). EDA is used for visualizing and studying data for the purpose of uncovering statistical regularities that might not be apparent otherwise. The analysis is carried out in terms of the formal notion of Weight of Evidence (WOE). As a result of this analysis, a novel theory in support of the use of inverse document frequency (idf) for document ranking is presented, and experimental evidence is given in favor of a modification of the classical idf formula motivated by the analysis. This approach is then extended to other sources of evidence commonly used for ranking in information retrieval systems
    Series
    The Kluwer international series on information retrieval; 7
    Source
    Advances in information retrieval: Recent research from the Center for Intelligent Information Retrieval. Ed.: W.B. Croft
  9. Rosemblat, G.; Graham, L.: Cross-language search in a monolingual health information system : flexible designs and lexical processes (2006) 0.08
    0.078905575 = product of:
      0.17753755 = sum of:
        0.08076138 = weight(_text_:line in 241) [ClassicSimilarity], result of:
          0.08076138 = score(doc=241,freq=2.0), product of:
            0.21724595 = queryWeight, product of:
              5.6078424 = idf(docFreq=440, maxDocs=44218)
              0.038739666 = queryNorm
            0.37175092 = fieldWeight in 241, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.6078424 = idf(docFreq=440, maxDocs=44218)
              0.046875 = fieldNorm(doc=241)
        0.013707667 = weight(_text_:information in 241) [ClassicSimilarity], result of:
          0.013707667 = score(doc=241,freq=6.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.20156369 = fieldWeight in 241, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=241)
        0.033231772 = weight(_text_:retrieval in 241) [ClassicSimilarity], result of:
          0.033231772 = score(doc=241,freq=4.0), product of:
            0.1171842 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.038739666 = queryNorm
            0.2835858 = fieldWeight in 241, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=241)
        0.049836725 = weight(_text_:techniques in 241) [ClassicSimilarity], result of:
          0.049836725 = score(doc=241,freq=2.0), product of:
            0.17065717 = queryWeight, product of:
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.038739666 = queryNorm
            0.2920283 = fieldWeight in 241, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.046875 = fieldNorm(doc=241)
      0.44444445 = coord(4/9)
    
    Abstract
    The predominance of English-only online health information poses a serious challenge to nonEnglish speakers. To overcome this barrier, we incorporated cross-language information retrieval (CLIR) techniques into a fully functional prototype. It supports Spanish language searches over an English data set using a Spanish-English bilingual term list (BTL). The modular design allows for system and BTL growth and takes advantage of English-system enhancements. Language-based design decisions and implications for integrating non-English components with the existing monolingual architecture are presented. Algorithmic and BTL improvements are used to bring CUR retrieval scores in line with the monolingual values. After validating these changes, we conducted a failure analysis and error categorization for the worst performing queries. We conclude with a comprehensive discussion and directions for future work.
  10. ¬The Eleventh Text Retrieval Conference, TREC 2002 (2003) 0.07
    0.07110572 = product of:
      0.15998787 = sum of:
        0.018276889 = weight(_text_:information in 4049) [ClassicSimilarity], result of:
          0.018276889 = score(doc=4049,freq=6.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.2687516 = fieldWeight in 4049, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=4049)
        0.05426726 = weight(_text_:retrieval in 4049) [ClassicSimilarity], result of:
          0.05426726 = score(doc=4049,freq=6.0), product of:
            0.1171842 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.038739666 = queryNorm
            0.46309367 = fieldWeight in 4049, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=4049)
        0.06644897 = weight(_text_:techniques in 4049) [ClassicSimilarity], result of:
          0.06644897 = score(doc=4049,freq=2.0), product of:
            0.17065717 = queryWeight, product of:
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.038739666 = queryNorm
            0.3893711 = fieldWeight in 4049, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.0625 = fieldNorm(doc=4049)
        0.02099476 = product of:
          0.04198952 = sum of:
            0.04198952 = weight(_text_:22 in 4049) [ClassicSimilarity], result of:
              0.04198952 = score(doc=4049,freq=2.0), product of:
                0.13565971 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038739666 = queryNorm
                0.30952093 = fieldWeight in 4049, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4049)
          0.5 = coord(1/2)
      0.44444445 = coord(4/9)
    
    Abstract
    Proceedings of the llth TREC-conference held in Gaithersburg, Maryland (USA), November 19-22, 2002. Aim of the conference was discussion an retrieval and related information-seeking tasks for large test collection. 93 research groups used different techniques, for information retrieval from the same large database. This procedure makes it possible to compare the results. The tasks are: Cross-language searching, filtering, interactive searching, searching for novelty, question answering, searching for video shots, and Web searching.
    Imprint
    Gaithersburg, MD : National Institute of Standards / Information Technology Laboratory
  11. Herrero-Solana, V.; Moya Anegón, F. de: Graphical Table of Contents (GTOC) for library collections : the application of UDC codes for the subject maps (2003) 0.07
    0.06876698 = product of:
      0.12378056 = sum of:
        0.053840924 = weight(_text_:line in 2758) [ClassicSimilarity], result of:
          0.053840924 = score(doc=2758,freq=2.0), product of:
            0.21724595 = queryWeight, product of:
              5.6078424 = idf(docFreq=440, maxDocs=44218)
              0.038739666 = queryNorm
            0.24783395 = fieldWeight in 2758, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.6078424 = idf(docFreq=440, maxDocs=44218)
              0.03125 = fieldNorm(doc=2758)
        0.010552166 = weight(_text_:information in 2758) [ClassicSimilarity], result of:
          0.010552166 = score(doc=2758,freq=8.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.1551638 = fieldWeight in 2758, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=2758)
        0.015665608 = weight(_text_:retrieval in 2758) [ClassicSimilarity], result of:
          0.015665608 = score(doc=2758,freq=2.0), product of:
            0.1171842 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.038739666 = queryNorm
            0.13368362 = fieldWeight in 2758, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=2758)
        0.033224486 = weight(_text_:techniques in 2758) [ClassicSimilarity], result of:
          0.033224486 = score(doc=2758,freq=2.0), product of:
            0.17065717 = queryWeight, product of:
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.038739666 = queryNorm
            0.19468555 = fieldWeight in 2758, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.03125 = fieldNorm(doc=2758)
        0.01049738 = product of:
          0.02099476 = sum of:
            0.02099476 = weight(_text_:22 in 2758) [ClassicSimilarity], result of:
              0.02099476 = score(doc=2758,freq=2.0), product of:
                0.13565971 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038739666 = queryNorm
                0.15476047 = fieldWeight in 2758, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2758)
          0.5 = coord(1/2)
      0.5555556 = coord(5/9)
    
    Abstract
    The representation of information contents by graphical maps is an extended ongoing research topic. In this paper we introduce the application of UDC codes for the subject maps development. We use the following graphic representation methodologies: 1) Multidimensional scaling (MDS), 2) Cluster analysis, 3) Neural networks (Self Organizing Map - SOM). Finally, we conclude about the application viability of every kind of map. 1. Introduction Advanced techniques for Information Retrieval (IR) currently make up one of the most active areas for research in the field of library and information science. New models representing document content are replacing the classic systems in which the search terms supplied by the user were compared against the indexing terms existing in the inverted files of a database. One of the topics most often studied in the last years is bibliographic browsing, a good complement to querying strategies. Since the 80's, many authors have treated this topic. For example, Ellis establishes that browsing is based an three different types of tasks: identification, familiarization and differentiation (Ellis, 1989). On the other hand, Cove indicates three different browsing types: searching browsing, general purpose browsing and serendipity browsing (Cove, 1988). Marcia Bates presents six different types (Bates, 1989), although the classification of Bawden is the one that really interests us: 1) similarity comparison, 2) structure driven, 3) global vision (Bawden, 1993). The global vision browsing implies the use of graphic representations, which we will call map displays, that allow the user to get a global idea of the nature and structure of the information in the database. In the 90's, several authors worked an this research line, developing different types of maps. One of the most active was Xia Lin what introduced the concept of Graphical Table of Contents (GTOC), comparing the maps to true table of contents based an graphic representations (Lin 1996). Lin applies the algorithm SOM to his own personal bibliography, analyzed in function of the words of the title and abstract fields, and represented in a two-dimensional map (Lin 1997). Later on, Lin applied this type of maps to create websites GTOCs, through a Java application.
    Date
    12. 9.2004 14:31:22
  12. Karamuftuoglu, M.: Need for a systemic theory of classification in information science (2007) 0.07
    0.06657348 = product of:
      0.14979032 = sum of:
        0.020938806 = weight(_text_:information in 615) [ClassicSimilarity], result of:
          0.020938806 = score(doc=615,freq=14.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.3078936 = fieldWeight in 615, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=615)
        0.040700447 = weight(_text_:retrieval in 615) [ClassicSimilarity], result of:
          0.040700447 = score(doc=615,freq=6.0), product of:
            0.1171842 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.038739666 = queryNorm
            0.34732026 = fieldWeight in 615, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=615)
        0.049836725 = weight(_text_:techniques in 615) [ClassicSimilarity], result of:
          0.049836725 = score(doc=615,freq=2.0), product of:
            0.17065717 = queryWeight, product of:
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.038739666 = queryNorm
            0.2920283 = fieldWeight in 615, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.046875 = fieldNorm(doc=615)
        0.03831434 = product of:
          0.07662868 = sum of:
            0.07662868 = weight(_text_:theories in 615) [ClassicSimilarity], result of:
              0.07662868 = score(doc=615,freq=2.0), product of:
                0.21161452 = queryWeight, product of:
                  5.4624767 = idf(docFreq=509, maxDocs=44218)
                  0.038739666 = queryNorm
                0.36211446 = fieldWeight in 615, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.4624767 = idf(docFreq=509, maxDocs=44218)
                  0.046875 = fieldNorm(doc=615)
          0.5 = coord(1/2)
      0.44444445 = coord(4/9)
    
    Abstract
    In the article, the author aims to clarify some of the issues surrounding the discussion regarding the usefulness of a substantive classification theory in information science (IS) by means of a broad perspective. By utilizing a concrete example from the High Accuracy Retrieval from Documents (HARD) track of a Text REtrieval Conference (TREC), the author suggests that the bag of words approach to information retrieval (IR) and techniques such as relevance feedback have significant limitations in expressing and resolving complex user information needs. He argues that a comprehensive analysis of information needs involves explicating often-implicit assumptions made by the authors of scholarly documents, as well as everyday texts such as news articles. He also argues that progress in IS can be furthered by developing general theories that are applicable to multiple domains. The concrete example of application of the domain-analytic approach to subject analysis in IS to the aesthetic evaluation of works of information arts is used to support this argument.
    Source
    Journal of the American Society for Information Science and Technology. 58(2007) no.13, S.1977-1987
  13. Gibb, F.: ¬The integration of information retrieval techniques within a software reuse environment (2000) 0.07
    0.065743536 = product of:
      0.1972306 = sum of:
        0.02611528 = weight(_text_:information in 2963) [ClassicSimilarity], result of:
          0.02611528 = score(doc=2963,freq=4.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.3840108 = fieldWeight in 2963, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=2963)
        0.054829627 = weight(_text_:retrieval in 2963) [ClassicSimilarity], result of:
          0.054829627 = score(doc=2963,freq=2.0), product of:
            0.1171842 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.038739666 = queryNorm
            0.46789268 = fieldWeight in 2963, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.109375 = fieldNorm(doc=2963)
        0.1162857 = weight(_text_:techniques in 2963) [ClassicSimilarity], result of:
          0.1162857 = score(doc=2963,freq=2.0), product of:
            0.17065717 = queryWeight, product of:
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.038739666 = queryNorm
            0.6813994 = fieldWeight in 2963, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.109375 = fieldNorm(doc=2963)
      0.33333334 = coord(3/9)
    
    Source
    Journal of information science. 26(2000) no.4, S.211-226
  14. Agosti, M.; Melucci, M.: Information retrieval techniques for the automatic construction of hypertext (2000) 0.07
    0.065743536 = product of:
      0.1972306 = sum of:
        0.02611528 = weight(_text_:information in 4671) [ClassicSimilarity], result of:
          0.02611528 = score(doc=4671,freq=4.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.3840108 = fieldWeight in 4671, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=4671)
        0.054829627 = weight(_text_:retrieval in 4671) [ClassicSimilarity], result of:
          0.054829627 = score(doc=4671,freq=2.0), product of:
            0.1171842 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.038739666 = queryNorm
            0.46789268 = fieldWeight in 4671, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.109375 = fieldNorm(doc=4671)
        0.1162857 = weight(_text_:techniques in 4671) [ClassicSimilarity], result of:
          0.1162857 = score(doc=4671,freq=2.0), product of:
            0.17065717 = queryWeight, product of:
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.038739666 = queryNorm
            0.6813994 = fieldWeight in 4671, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.109375 = fieldNorm(doc=4671)
      0.33333334 = coord(3/9)
    
    Source
    Encyclopedia of library and information science. Vol.66, [=Suppl.29]
  15. Mengle, S.; Goharian, N.: Passage detection using text classification (2009) 0.06
    0.06312523 = product of:
      0.14203176 = sum of:
        0.013190207 = weight(_text_:information in 2765) [ClassicSimilarity], result of:
          0.013190207 = score(doc=2765,freq=8.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.19395474 = fieldWeight in 2765, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2765)
        0.043786705 = weight(_text_:retrieval in 2765) [ClassicSimilarity], result of:
          0.043786705 = score(doc=2765,freq=10.0), product of:
            0.1171842 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.038739666 = queryNorm
            0.37365708 = fieldWeight in 2765, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2765)
        0.07193312 = weight(_text_:techniques in 2765) [ClassicSimilarity], result of:
          0.07193312 = score(doc=2765,freq=6.0), product of:
            0.17065717 = queryWeight, product of:
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.038739666 = queryNorm
            0.42150658 = fieldWeight in 2765, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2765)
        0.013121725 = product of:
          0.02624345 = sum of:
            0.02624345 = weight(_text_:22 in 2765) [ClassicSimilarity], result of:
              0.02624345 = score(doc=2765,freq=2.0), product of:
                0.13565971 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038739666 = queryNorm
                0.19345059 = fieldWeight in 2765, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2765)
          0.5 = coord(1/2)
      0.44444445 = coord(4/9)
    
    Abstract
    Passages can be hidden within a text to circumvent their disallowed transfer. Such release of compartmentalized information is of concern to all corporate and governmental organizations. Passage retrieval is well studied; we posit, however, that passage detection is not. Passage retrieval is the determination of the degree of relevance of blocks of text, namely passages, comprising a document. Rather than determining the relevance of a document in its entirety, passage retrieval determines the relevance of the individual passages. As such, modified traditional information-retrieval techniques compare terms found in user queries with the individual passages to determine a similarity score for passages of interest. In passage detection, passages are classified into predetermined categories. More often than not, passage detection techniques are deployed to detect hidden paragraphs in documents. That is, to hide information, documents are injected with hidden text into passages. Rather than matching query terms against passages to determine their relevance, using text-mining techniques, the passages are classified. Those documents with hidden passages are defined as infected. Thus, simply stated, passage retrieval is the search for passages relevant to a user query, while passage detection is the classification of passages. That is, in passage detection, passages are labeled with one or more categories from a set of predetermined categories. We present a keyword-based dynamic passage approach (KDP) and demonstrate that KDP outperforms statistically significantly (99% confidence) the other document-splitting approaches by 12% to 18% in the passage detection and passage category-prediction tasks. Furthermore, we evaluate the effects of the feature selection, passage length, ambiguous passages, and finally training-data category distribution on passage-detection accuracy.
    Date
    22. 3.2009 19:14:43
    Source
    Journal of the American Society for Information Science and Technology. 60(2009) no.4, S.814-825
  16. Aringhieri, R.; Damiani, E.; De Capitani di Vimercati, S.; Paraboschi, S.; Samarati, P.: Fuzzy techniques for trust and reputation management in anonymous peer-to-peer systems (2006) 0.06
    0.061898623 = product of:
      0.1392719 = sum of:
        0.013707667 = weight(_text_:information in 5279) [ClassicSimilarity], result of:
          0.013707667 = score(doc=5279,freq=6.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.20156369 = fieldWeight in 5279, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=5279)
        0.023498412 = weight(_text_:retrieval in 5279) [ClassicSimilarity], result of:
          0.023498412 = score(doc=5279,freq=2.0), product of:
            0.1171842 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.038739666 = queryNorm
            0.20052543 = fieldWeight in 5279, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=5279)
        0.086319745 = weight(_text_:techniques in 5279) [ClassicSimilarity], result of:
          0.086319745 = score(doc=5279,freq=6.0), product of:
            0.17065717 = queryWeight, product of:
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.038739666 = queryNorm
            0.5058079 = fieldWeight in 5279, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.046875 = fieldNorm(doc=5279)
        0.01574607 = product of:
          0.03149214 = sum of:
            0.03149214 = weight(_text_:22 in 5279) [ClassicSimilarity], result of:
              0.03149214 = score(doc=5279,freq=2.0), product of:
                0.13565971 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038739666 = queryNorm
                0.23214069 = fieldWeight in 5279, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5279)
          0.5 = coord(1/2)
      0.44444445 = coord(4/9)
    
    Abstract
    Peer-to-peer (P2P) applications are rapidly gaining acceptance among users of Internet-based services, especially because of their capability of exchanging resources while preserving the anonymity of both requesters and providers. However, concerns have been raised about the possibility that malicious users can exploit the network to spread tampered-with resources (e.g., malicious programs and viruses). A considerable amount of research has thus focused on the development of trust and reputation models in P2P networks. In this article, we propose to use fuzzy techniques in the design of reputation systems based on collecting and aggregating peers' opinions. Fuzzy techniques are used in the evaluation and synthesis of all the opinions expressed by peers. The behavior of the proposed system is described by comparison with probabilistic approaches.
    Date
    22. 7.2006 17:06:18
    Footnote
    Beitrag in einer Special Topic Section on Soft Approaches to Information Retrieval and Information Access on the Web
    Source
    Journal of the American Society for Information Science and Technology. 57(2006) no.4, S.528-537
  17. Marion, L.S.; McCain, K.W.: Contrasting views of software engineering journals : author cocitation choices and indexer vocabulary assignments (2001) 0.06
    0.06121807 = product of:
      0.13774066 = sum of:
        0.067301154 = weight(_text_:line in 5767) [ClassicSimilarity], result of:
          0.067301154 = score(doc=5767,freq=2.0), product of:
            0.21724595 = queryWeight, product of:
              5.6078424 = idf(docFreq=440, maxDocs=44218)
              0.038739666 = queryNorm
            0.30979243 = fieldWeight in 5767, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.6078424 = idf(docFreq=440, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5767)
        0.009326885 = weight(_text_:information in 5767) [ClassicSimilarity], result of:
          0.009326885 = score(doc=5767,freq=4.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.13714671 = fieldWeight in 5767, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5767)
        0.01958201 = weight(_text_:retrieval in 5767) [ClassicSimilarity], result of:
          0.01958201 = score(doc=5767,freq=2.0), product of:
            0.1171842 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.038739666 = queryNorm
            0.16710453 = fieldWeight in 5767, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5767)
        0.04153061 = weight(_text_:techniques in 5767) [ClassicSimilarity], result of:
          0.04153061 = score(doc=5767,freq=2.0), product of:
            0.17065717 = queryWeight, product of:
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.038739666 = queryNorm
            0.24335694 = fieldWeight in 5767, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5767)
      0.44444445 = coord(4/9)
    
    Abstract
    We explore the intellectual subject structure and research themes in software engineering through the identification and analysis of a core journal literature. We examine this literature via two expert perspectives: that of the author, who identified significant work by citing it (journal cocitation analysis), and that of the professional indexer, who tags published work with subject terms to facilitate retrieval from a bibliographic database (subject profile analysis). The data sources are SCISEARCH (the on-line version of Science Citation Index), and INSPEC (a database covering software engineering, computer science, and information systems). We use data visualization tools (cluster analysis, multidimensional scaling, and PFNets) to show the "intellectual maps" of software engineering. Cocitation and subject profile analyses demonstrate that software engineering is a distinct interdisciplinary field, valuing practical and applied aspects, and spanning a subject continuum from "programming-in-the-smalI" to "programming-in-the-large." This continuum mirrors the software development life cycle by taking the operating system or major application from initial programming through project management, implementation, and maintenance. Object orientation is an integral but distinct subject area in software engineering. Key differences are the importance of management and programming: (1) cocitation analysis emphasizes project management and systems development; (2) programming techniques/languages are more influential in subject profiles; (3) cocitation profiles place object-oriented journals separately and centrally while the subject profile analysis locates these journals with the programming/languages group
    Source
    Journal of the American Society for Information Science and technology. 52(2001) no.4, S.297-308
  18. Freitas-Junior, H.R.; Ribeiro-Neto, B.A.; Freitas-Vale, R. de; Laender, A.H.F.; Lima, L.R.S. de: Categorization-driven cross-language retrieval of medical information (2006) 0.06
    0.060433533 = product of:
      0.13597545 = sum of:
        0.01615464 = weight(_text_:information in 5282) [ClassicSimilarity], result of:
          0.01615464 = score(doc=5282,freq=12.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.23754507 = fieldWeight in 5282, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5282)
        0.047965933 = weight(_text_:retrieval in 5282) [ClassicSimilarity], result of:
          0.047965933 = score(doc=5282,freq=12.0), product of:
            0.1171842 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.038739666 = queryNorm
            0.40932083 = fieldWeight in 5282, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5282)
        0.058733147 = weight(_text_:techniques in 5282) [ClassicSimilarity], result of:
          0.058733147 = score(doc=5282,freq=4.0), product of:
            0.17065717 = queryWeight, product of:
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.038739666 = queryNorm
            0.34415868 = fieldWeight in 5282, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5282)
        0.013121725 = product of:
          0.02624345 = sum of:
            0.02624345 = weight(_text_:22 in 5282) [ClassicSimilarity], result of:
              0.02624345 = score(doc=5282,freq=2.0), product of:
                0.13565971 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038739666 = queryNorm
                0.19345059 = fieldWeight in 5282, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5282)
          0.5 = coord(1/2)
      0.44444445 = coord(4/9)
    
    Abstract
    The Web has become a large repository of documents (or pages) written in many different languages. In this context, traditional information retrieval (IR) techniques cannot be used whenever the user query and the documents being retrieved are in different languages. To address this problem, new cross-language information retrieval (CLIR) techniques have been proposed. In this work, we describe a method for cross-language retrieval of medical information. This method combines query terms and related medical concepts obtained automatically through a categorization procedure. The medical concepts are used to create a linguistic abstraction that allows retrieval of information in a language-independent way, minimizing linguistic problems such as polysemy. To evaluate our method, we carried out experiments using the OHSUMED test collection, whose documents are written in English, with queries expressed in Portuguese, Spanish, and French. The results indicate that our cross-language retrieval method is as effective as a standard vector space model algorithm operating on queries and documents in the same language. Further, our results are better than previous results in the literature.
    Date
    22. 7.2006 16:46:36
    Source
    Journal of the American Society for Information Science and Technology. 57(2006) no.4, S.501-510
  19. Information visualization in data mining and knowledge discovery (2002) 0.06
    0.060118683 = product of:
      0.108213626 = sum of:
        0.026920462 = weight(_text_:line in 1789) [ClassicSimilarity], result of:
          0.026920462 = score(doc=1789,freq=2.0), product of:
            0.21724595 = queryWeight, product of:
              5.6078424 = idf(docFreq=440, maxDocs=44218)
              0.038739666 = queryNorm
            0.123916976 = fieldWeight in 1789, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.6078424 = idf(docFreq=440, maxDocs=44218)
              0.015625 = fieldNorm(doc=1789)
        0.009870647 = weight(_text_:information in 1789) [ClassicSimilarity], result of:
          0.009870647 = score(doc=1789,freq=28.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.14514244 = fieldWeight in 1789, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.015625 = fieldNorm(doc=1789)
        0.011077258 = weight(_text_:retrieval in 1789) [ClassicSimilarity], result of:
          0.011077258 = score(doc=1789,freq=4.0), product of:
            0.1171842 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.038739666 = queryNorm
            0.09452859 = fieldWeight in 1789, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.015625 = fieldNorm(doc=1789)
        0.055096574 = weight(_text_:techniques in 1789) [ClassicSimilarity], result of:
          0.055096574 = score(doc=1789,freq=22.0), product of:
            0.17065717 = queryWeight, product of:
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.038739666 = queryNorm
            0.32284945 = fieldWeight in 1789, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.015625 = fieldNorm(doc=1789)
        0.00524869 = product of:
          0.01049738 = sum of:
            0.01049738 = weight(_text_:22 in 1789) [ClassicSimilarity], result of:
              0.01049738 = score(doc=1789,freq=2.0), product of:
                0.13565971 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038739666 = queryNorm
                0.07738023 = fieldWeight in 1789, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.015625 = fieldNorm(doc=1789)
          0.5 = coord(1/2)
      0.5555556 = coord(5/9)
    
    Date
    23. 3.2008 19:10:22
    Footnote
    Rez. in: JASIST 54(2003) no.9, S.905-906 (C.A. Badurek): "Visual approaches for knowledge discovery in very large databases are a prime research need for information scientists focused an extracting meaningful information from the ever growing stores of data from a variety of domains, including business, the geosciences, and satellite and medical imagery. This work presents a summary of research efforts in the fields of data mining, knowledge discovery, and data visualization with the goal of aiding the integration of research approaches and techniques from these major fields. The editors, leading computer scientists from academia and industry, present a collection of 32 papers from contributors who are incorporating visualization and data mining techniques through academic research as well application development in industry and government agencies. Information Visualization focuses upon techniques to enhance the natural abilities of humans to visually understand data, in particular, large-scale data sets. It is primarily concerned with developing interactive graphical representations to enable users to more intuitively make sense of multidimensional data as part of the data exploration process. It includes research from computer science, psychology, human-computer interaction, statistics, and information science. Knowledge Discovery in Databases (KDD) most often refers to the process of mining databases for previously unknown patterns and trends in data. Data mining refers to the particular computational methods or algorithms used in this process. The data mining research field is most related to computational advances in database theory, artificial intelligence and machine learning. This work compiles research summaries from these main research areas in order to provide "a reference work containing the collection of thoughts and ideas of noted researchers from the fields of data mining and data visualization" (p. 8). It addresses these areas in three main sections: the first an data visualization, the second an KDD and model visualization, and the last an using visualization in the knowledge discovery process. The seven chapters of Part One focus upon methodologies and successful techniques from the field of Data Visualization. Hoffman and Grinstein (Chapter 2) give a particularly good overview of the field of data visualization and its potential application to data mining. An introduction to the terminology of data visualization, relation to perceptual and cognitive science, and discussion of the major visualization display techniques are presented. Discussion and illustration explain the usefulness and proper context of such data visualization techniques as scatter plots, 2D and 3D isosurfaces, glyphs, parallel coordinates, and radial coordinate visualizations. Remaining chapters present the need for standardization of visualization methods, discussion of user requirements in the development of tools, and examples of using information visualization in addressing research problems.
    In 13 chapters, Part Two provides an introduction to KDD, an overview of data mining techniques, and examples of the usefulness of data model visualizations. The importance of visualization throughout the KDD process is stressed in many of the chapters. In particular, the need for measures of visualization effectiveness, benchmarking for identifying best practices, and the use of standardized sample data sets is convincingly presented. Many of the important data mining approaches are discussed in this complementary context. Cluster and outlier detection, classification techniques, and rule discovery algorithms are presented as the basic techniques common to the KDD process. The potential effectiveness of using visualization in the data modeling process are illustrated in chapters focused an using visualization for helping users understand the KDD process, ask questions and form hypotheses about their data, and evaluate the accuracy and veracity of their results. The 11 chapters of Part Three provide an overview of the KDD process and successful approaches to integrating KDD, data mining, and visualization in complementary domains. Rhodes (Chapter 21) begins this section with an excellent overview of the relation between the KDD process and data mining techniques. He states that the "primary goals of data mining are to describe the existing data and to predict the behavior or characteristics of future data of the same type" (p. 281). These goals are met by data mining tasks such as classification, regression, clustering, summarization, dependency modeling, and change or deviation detection. Subsequent chapters demonstrate how visualization can aid users in the interactive process of knowledge discovery by graphically representing the results from these iterative tasks. Finally, examples of the usefulness of integrating visualization and data mining tools in the domain of business, imagery and text mining, and massive data sets are provided. This text concludes with a thorough and useful 17-page index and lengthy yet integrating 17-page summary of the academic and industrial backgrounds of the contributing authors. A 16-page set of color inserts provide a better representation of the visualizations discussed, and a URL provided suggests that readers may view all the book's figures in color on-line, although as of this submission date it only provides access to a summary of the book and its contents. The overall contribution of this work is its focus an bridging two distinct areas of research, making it a valuable addition to the Morgan Kaufmann Series in Database Management Systems. The editors of this text have met their main goal of providing the first textbook integrating knowledge discovery, data mining, and visualization. Although it contributes greatly to our under- standing of the development and current state of the field, a major weakness of this text is that there is no concluding chapter to discuss the contributions of the sum of these contributed papers or give direction to possible future areas of research. "Integration of expertise between two different disciplines is a difficult process of communication and reeducation. Integrating data mining and visualization is particularly complex because each of these fields in itself must draw an a wide range of research experience" (p. 300). Although this work contributes to the crossdisciplinary communication needed to advance visualization in KDD, a more formal call for an interdisciplinary research agenda in a concluding chapter would have provided a more satisfying conclusion to a very good introductory text.
    With contributors almost exclusively from the computer science field, the intended audience of this work is heavily slanted towards a computer science perspective. However, it is highly readable and provides introductory material that would be useful to information scientists from a variety of domains. Yet, much interesting work in information visualization from other fields could have been included giving the work more of an interdisciplinary perspective to complement their goals of integrating work in this area. Unfortunately, many of the application chapters are these, shallow, and lack complementary illustrations of visualization techniques or user interfaces used. However, they do provide insight into the many applications being developed in this rapidly expanding field. The authors have successfully put together a highly useful reference text for the data mining and information visualization communities. Those interested in a good introduction and overview of complementary research areas in these fields will be satisfied with this collection of papers. The focus upon integrating data visualization with data mining complements texts in each of these fields, such as Advances in Knowledge Discovery and Data Mining (Fayyad et al., MIT Press) and Readings in Information Visualization: Using Vision to Think (Card et. al., Morgan Kauffman). This unique work is a good starting point for future interaction between researchers in the fields of data visualization and data mining and makes a good accompaniment for a course focused an integrating these areas or to the main reference texts in these fields."
    LCSH
    Information visualization
    RSWK
    Information Retrieval (BVB)
    Subject
    Information Retrieval (BVB)
    Information visualization
  20. Soft computing in information retrieval : techniques and applications (2000) 0.06
    0.0592466 = product of:
      0.1777398 = sum of:
        0.021104332 = weight(_text_:information in 4947) [ClassicSimilarity], result of:
          0.021104332 = score(doc=4947,freq=8.0), product of:
            0.06800663 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.038739666 = queryNorm
            0.3103276 = fieldWeight in 4947, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=4947)
        0.06266243 = weight(_text_:retrieval in 4947) [ClassicSimilarity], result of:
          0.06266243 = score(doc=4947,freq=8.0), product of:
            0.1171842 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.038739666 = queryNorm
            0.5347345 = fieldWeight in 4947, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=4947)
        0.09397303 = weight(_text_:techniques in 4947) [ClassicSimilarity], result of:
          0.09397303 = score(doc=4947,freq=4.0), product of:
            0.17065717 = queryWeight, product of:
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.038739666 = queryNorm
            0.5506539 = fieldWeight in 4947, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.405231 = idf(docFreq=1467, maxDocs=44218)
              0.0625 = fieldNorm(doc=4947)
      0.33333334 = coord(3/9)
    
    Abstract
    Presented are a number of advanced models for the representation and retrieval of information originating from the application of soft computing techniques to information retrieval. The book is a collection of articles from some of the most outstanding and well known researchers in the area of information retrieval

Languages

Types

  • a 6085
  • m 748
  • el 390
  • s 251
  • x 130
  • b 38
  • r 33
  • i 28
  • n 14
  • p 1
  • More… Less…

Themes

Subjects

Classifications