Search (8177 results, page 1 of 409)

  • × year_i:[2000 TO 2010}
  1. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.44
    0.43542933 = product of:
      0.7464503 = sum of:
        0.054483652 = product of:
          0.16345096 = sum of:
            0.16345096 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
              0.16345096 = score(doc=562,freq=2.0), product of:
                0.29082868 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03430388 = queryNorm
                0.56201804 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
        0.024219744 = weight(_text_:web in 562) [ClassicSimilarity], result of:
          0.024219744 = score(doc=562,freq=2.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.21634221 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.16345096 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.16345096 = score(doc=562,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.16345096 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.16345096 = score(doc=562,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.16345096 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.16345096 = score(doc=562,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.16345096 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.16345096 = score(doc=562,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.013943106 = product of:
          0.027886212 = sum of:
            0.027886212 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
              0.027886212 = score(doc=562,freq=2.0), product of:
                0.120126344 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03430388 = queryNorm
                0.23214069 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.5 = coord(1/2)
      0.5833333 = coord(7/12)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  2. Mas, S.; Marleau, Y.: Proposition of a faceted classification model to support corporate information organization and digital records management (2009) 0.43
    0.4321077 = product of:
      0.74075603 = sum of:
        0.054483652 = product of:
          0.16345096 = sum of:
            0.16345096 = weight(_text_:3a in 2918) [ClassicSimilarity], result of:
              0.16345096 = score(doc=2918,freq=2.0), product of:
                0.29082868 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03430388 = queryNorm
                0.56201804 = fieldWeight in 2918, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2918)
          0.33333334 = coord(1/3)
        0.16345096 = weight(_text_:2f in 2918) [ClassicSimilarity], result of:
          0.16345096 = score(doc=2918,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.56201804 = fieldWeight in 2918, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=2918)
        0.009910721 = weight(_text_:information in 2918) [ClassicSimilarity], result of:
          0.009910721 = score(doc=2918,freq=4.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.16457605 = fieldWeight in 2918, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2918)
        0.16345096 = weight(_text_:2f in 2918) [ClassicSimilarity], result of:
          0.16345096 = score(doc=2918,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.56201804 = fieldWeight in 2918, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=2918)
        0.16345096 = weight(_text_:2f in 2918) [ClassicSimilarity], result of:
          0.16345096 = score(doc=2918,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.56201804 = fieldWeight in 2918, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=2918)
        0.02255783 = weight(_text_:system in 2918) [ClassicSimilarity], result of:
          0.02255783 = score(doc=2918,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.20878783 = fieldWeight in 2918, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.046875 = fieldNorm(doc=2918)
        0.16345096 = weight(_text_:2f in 2918) [ClassicSimilarity], result of:
          0.16345096 = score(doc=2918,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.56201804 = fieldWeight in 2918, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=2918)
      0.5833333 = coord(7/12)
    
    Abstract
    The employees of an organization often use a personal hierarchical classification scheme to organize digital documents that are stored on their own workstations. As this may make it hard for other employees to retrieve these documents, there is a risk that the organization will lose track of needed documentation. Furthermore, the inherent boundaries of such a hierarchical structure require making arbitrary decisions about which specific criteria the classification will b.e based on (for instance, the administrative activity or the document type, although a document can have several attributes and require classification in several classes).A faceted classification model to support corporate information organization is proposed. Partially based on Ranganathan's facets theory, this model aims not only to standardize the organization of digital documents, but also to simplify the management of a document throughout its life cycle for both individuals and organizations, while ensuring compliance to regulatory and policy requirements.
    Footnote
    Vgl.: http://ieeexplore.ieee.org/Xplore/login.jsp?reload=true&url=http%3A%2F%2Fieeexplore.ieee.org%2Fiel5%2F4755313%2F4755314%2F04755480.pdf%3Farnumber%3D4755480&authDecision=-203.
    Source
    System Sciences, 2009. HICSS '09. 42nd Hawaii International Conference
  3. Schrodt, R.: Tiefen und Untiefen im wissenschaftlichen Sprachgebrauch (2008) 0.39
    0.39349306 = product of:
      0.9443833 = sum of:
        0.072644874 = product of:
          0.21793462 = sum of:
            0.21793462 = weight(_text_:3a in 140) [ClassicSimilarity], result of:
              0.21793462 = score(doc=140,freq=2.0), product of:
                0.29082868 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03430388 = queryNorm
                0.7493574 = fieldWeight in 140, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0625 = fieldNorm(doc=140)
          0.33333334 = coord(1/3)
        0.21793462 = weight(_text_:2f in 140) [ClassicSimilarity], result of:
          0.21793462 = score(doc=140,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.7493574 = fieldWeight in 140, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0625 = fieldNorm(doc=140)
        0.21793462 = weight(_text_:2f in 140) [ClassicSimilarity], result of:
          0.21793462 = score(doc=140,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.7493574 = fieldWeight in 140, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0625 = fieldNorm(doc=140)
        0.21793462 = weight(_text_:2f in 140) [ClassicSimilarity], result of:
          0.21793462 = score(doc=140,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.7493574 = fieldWeight in 140, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0625 = fieldNorm(doc=140)
        0.21793462 = weight(_text_:2f in 140) [ClassicSimilarity], result of:
          0.21793462 = score(doc=140,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.7493574 = fieldWeight in 140, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0625 = fieldNorm(doc=140)
      0.41666666 = coord(5/12)
    
    Content
    Vgl. auch: https://studylibde.com/doc/13053640/richard-schrodt. Vgl. auch: http%3A%2F%2Fwww.univie.ac.at%2FGermanistik%2Fschrodt%2Fvorlesung%2Fwissenschaftssprache.doc&usg=AOvVaw1lDLDR6NFf1W0-oC9mEUJf.
  4. Stojanovic, N.: Ontology-based Information Retrieval : methods and tools for cooperative query answering (2005) 0.36
    0.35672647 = product of:
      0.5350897 = sum of:
        0.036322437 = product of:
          0.10896731 = sum of:
            0.10896731 = weight(_text_:3a in 701) [ClassicSimilarity], result of:
              0.10896731 = score(doc=701,freq=2.0), product of:
                0.29082868 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03430388 = queryNorm
                0.3746787 = fieldWeight in 701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
          0.33333334 = coord(1/3)
        0.022834593 = weight(_text_:web in 701) [ClassicSimilarity], result of:
          0.022834593 = score(doc=701,freq=4.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.2039694 = fieldWeight in 701, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.10896731 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.10896731 = score(doc=701,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.014015876 = weight(_text_:information in 701) [ClassicSimilarity], result of:
          0.014015876 = score(doc=701,freq=18.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.23274568 = fieldWeight in 701, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.10896731 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.10896731 = score(doc=701,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.10896731 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.10896731 = score(doc=701,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.026047539 = weight(_text_:system in 701) [ClassicSimilarity], result of:
          0.026047539 = score(doc=701,freq=6.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.24108742 = fieldWeight in 701, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.10896731 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.10896731 = score(doc=701,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
      0.6666667 = coord(8/12)
    
    Abstract
    By the explosion of possibilities for a ubiquitous content production, the information overload problem reaches the level of complexity which cannot be managed by traditional modelling approaches anymore. Due to their pure syntactical nature traditional information retrieval approaches did not succeed in treating content itself (i.e. its meaning, and not its representation). This leads to a very low usefulness of the results of a retrieval process for a user's task at hand. In the last ten years ontologies have been emerged from an interesting conceptualisation paradigm to a very promising (semantic) modelling technology, especially in the context of the Semantic Web. From the information retrieval point of view, ontologies enable a machine-understandable form of content description, such that the retrieval process can be driven by the meaning of the content. However, the very ambiguous nature of the retrieval process in which a user, due to the unfamiliarity with the underlying repository and/or query syntax, just approximates his information need in a query, implies a necessity to include the user in the retrieval process more actively in order to close the gap between the meaning of the content and the meaning of a user's query (i.e. his information need). This thesis lays foundation for such an ontology-based interactive retrieval process, in which the retrieval system interacts with a user in order to conceptually interpret the meaning of his query, whereas the underlying domain ontology drives the conceptualisation process. In that way the retrieval process evolves from a query evaluation process into a highly interactive cooperation between a user and the retrieval system, in which the system tries to anticipate the user's information need and to deliver the relevant content proactively. Moreover, the notion of content relevance for a user's query evolves from a content dependent artefact to the multidimensional context-dependent structure, strongly influenced by the user's preferences. This cooperation process is realized as the so-called Librarian Agent Query Refinement Process. In order to clarify the impact of an ontology on the retrieval process (regarding its complexity and quality), a set of methods and tools for different levels of content and query formalisation is developed, ranging from pure ontology-based inferencing to keyword-based querying in which semantics automatically emerges from the results. Our evaluation studies have shown that the possibilities to conceptualize a user's information need in the right manner and to interpret the retrieval results accordingly are key issues for realizing much more meaningful information retrieval systems.
    Content
    Vgl.: http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F1627&ei=tAtYUYrBNoHKtQb3l4GYBw&usg=AFQjCNHeaxKkKU3-u54LWxMNYGXaaDLCGw&sig2=8WykXWQoDKjDSdGtAakH2Q&bvm=bv.44442042,d.Yms.
    Theme
    Semantic Web
  5. Vetere, G.; Lenzerini, M.: Models for semantic interoperability in service-oriented architectures (2005) 0.34
    0.3443064 = product of:
      0.82633543 = sum of:
        0.06356427 = product of:
          0.1906928 = sum of:
            0.1906928 = weight(_text_:3a in 306) [ClassicSimilarity], result of:
              0.1906928 = score(doc=306,freq=2.0), product of:
                0.29082868 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03430388 = queryNorm
                0.65568775 = fieldWeight in 306, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=306)
          0.33333334 = coord(1/3)
        0.1906928 = weight(_text_:2f in 306) [ClassicSimilarity], result of:
          0.1906928 = score(doc=306,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.65568775 = fieldWeight in 306, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0546875 = fieldNorm(doc=306)
        0.1906928 = weight(_text_:2f in 306) [ClassicSimilarity], result of:
          0.1906928 = score(doc=306,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.65568775 = fieldWeight in 306, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0546875 = fieldNorm(doc=306)
        0.1906928 = weight(_text_:2f in 306) [ClassicSimilarity], result of:
          0.1906928 = score(doc=306,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.65568775 = fieldWeight in 306, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0546875 = fieldNorm(doc=306)
        0.1906928 = weight(_text_:2f in 306) [ClassicSimilarity], result of:
          0.1906928 = score(doc=306,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.65568775 = fieldWeight in 306, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0546875 = fieldNorm(doc=306)
      0.41666666 = coord(5/12)
    
    Content
    Vgl.: http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=5386707&url=http%3A%2F%2Fieeexplore.ieee.org%2Fxpls%2Fabs_all.jsp%3Farnumber%3D5386707.
  6. Donsbach, W.: Wahrheit in den Medien : über den Sinn eines methodischen Objektivitätsbegriffes (2001) 0.30
    0.29803976 = product of:
      0.5960795 = sum of:
        0.04540305 = product of:
          0.13620915 = sum of:
            0.13620915 = weight(_text_:3a in 5895) [ClassicSimilarity], result of:
              0.13620915 = score(doc=5895,freq=2.0), product of:
                0.29082868 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03430388 = queryNorm
                0.46834838 = fieldWeight in 5895, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5895)
          0.33333334 = coord(1/3)
        0.13620915 = weight(_text_:2f in 5895) [ClassicSimilarity], result of:
          0.13620915 = score(doc=5895,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.46834838 = fieldWeight in 5895, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5895)
        0.0058399485 = weight(_text_:information in 5895) [ClassicSimilarity], result of:
          0.0058399485 = score(doc=5895,freq=2.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.09697737 = fieldWeight in 5895, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5895)
        0.13620915 = weight(_text_:2f in 5895) [ClassicSimilarity], result of:
          0.13620915 = score(doc=5895,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.46834838 = fieldWeight in 5895, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5895)
        0.13620915 = weight(_text_:2f in 5895) [ClassicSimilarity], result of:
          0.13620915 = score(doc=5895,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.46834838 = fieldWeight in 5895, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5895)
        0.13620915 = weight(_text_:2f in 5895) [ClassicSimilarity], result of:
          0.13620915 = score(doc=5895,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.46834838 = fieldWeight in 5895, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5895)
      0.5 = coord(6/12)
    
    Source
    Politische Meinung. 381(2001) Nr.1, S.65-74 [https%3A%2F%2Fwww.dgfe.de%2Ffileadmin%2FOrdnerRedakteure%2FSektionen%2FSek02_AEW%2FKWF%2FPublikationen_Reihe_1989-2003%2FBand_17%2FBd_17_1994_355-406_A.pdf&usg=AOvVaw2KcbRsHy5UQ9QRIUyuOLNi]
    Theme
    Information
  7. Knowledge organization, information systems and other essays : Professor A. Neelameghan festschrift (2006) 0.21
    0.21456532 = product of:
      0.64369595 = sum of:
        0.011679897 = weight(_text_:information in 6332) [ClassicSimilarity], result of:
          0.011679897 = score(doc=6332,freq=2.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.19395474 = fieldWeight in 6332, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=6332)
        0.41569185 = weight(_text_:wissensbasiertes in 6332) [ClassicSimilarity], result of:
          0.41569185 = score(doc=6332,freq=4.0), product of:
            0.30209768 = queryWeight, product of:
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.03430388 = queryNorm
            1.376018 = fieldWeight in 6332, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.078125 = fieldNorm(doc=6332)
        0.053169318 = weight(_text_:system in 6332) [ClassicSimilarity], result of:
          0.053169318 = score(doc=6332,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.49211764 = fieldWeight in 6332, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.078125 = fieldNorm(doc=6332)
        0.16315487 = product of:
          0.32630974 = sum of:
            0.32630974 = weight(_text_:aufsatzsammlung in 6332) [ClassicSimilarity], result of:
              0.32630974 = score(doc=6332,freq=8.0), product of:
                0.2250708 = queryWeight, product of:
                  6.5610886 = idf(docFreq=169, maxDocs=44218)
                  0.03430388 = queryNorm
                1.4498093 = fieldWeight in 6332, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  6.5610886 = idf(docFreq=169, maxDocs=44218)
                  0.078125 = fieldNorm(doc=6332)
          0.5 = coord(1/2)
      0.33333334 = coord(4/12)
    
    RSWK
    Indien / Wissensbasiertes System / Aufsatzsammlung
    Indien / Informations- und Dokumentationswissenschaft / Aufsatzsammlung
    Subject
    Indien / Wissensbasiertes System / Aufsatzsammlung
    Indien / Informations- und Dokumentationswissenschaft / Aufsatzsammlung
  8. Handbook on ontologies (2004) 0.16
    0.15931618 = product of:
      0.38235885 = sum of:
        0.03495819 = weight(_text_:web in 1952) [ClassicSimilarity], result of:
          0.03495819 = score(doc=1952,freq=6.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.3122631 = fieldWeight in 1952, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1952)
        0.013058522 = weight(_text_:information in 1952) [ClassicSimilarity], result of:
          0.013058522 = score(doc=1952,freq=10.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.21684799 = fieldWeight in 1952, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1952)
        0.20784593 = weight(_text_:wissensbasiertes in 1952) [ClassicSimilarity], result of:
          0.20784593 = score(doc=1952,freq=4.0), product of:
            0.30209768 = queryWeight, product of:
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.03430388 = queryNorm
            0.688009 = fieldWeight in 1952, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1952)
        0.026584659 = weight(_text_:system in 1952) [ClassicSimilarity], result of:
          0.026584659 = score(doc=1952,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.24605882 = fieldWeight in 1952, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1952)
        0.099911556 = product of:
          0.19982311 = sum of:
            0.19982311 = weight(_text_:aufsatzsammlung in 1952) [ClassicSimilarity], result of:
              0.19982311 = score(doc=1952,freq=12.0), product of:
                0.2250708 = queryWeight, product of:
                  6.5610886 = idf(docFreq=169, maxDocs=44218)
                  0.03430388 = queryNorm
                0.88782334 = fieldWeight in 1952, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  6.5610886 = idf(docFreq=169, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1952)
          0.5 = coord(1/2)
      0.41666666 = coord(5/12)
    
    LCSH
    Knowledge representation (Information theory)
    Conceptual structures (Information theory)
    RSWK
    Informationssystem / Wissenstechnik / Ontologie <Wissensverarbeitung> / Aufsatzsammlung
    Semantic Web / Wissensbasiertes System / Ontologie <Wissensverarbeitung> / Aufsatzsammlung
    Informationssystem / Konzeptionelle Modellierung / Ontologie <Wissensverarbeitung> / Aufsatzsammlung
    Series
    International handbook on information systems
    Subject
    Informationssystem / Wissenstechnik / Ontologie <Wissensverarbeitung> / Aufsatzsammlung
    Semantic Web / Wissensbasiertes System / Ontologie <Wissensverarbeitung> / Aufsatzsammlung
    Informationssystem / Konzeptionelle Modellierung / Ontologie <Wissensverarbeitung> / Aufsatzsammlung
    Knowledge representation (Information theory)
    Conceptual structures (Information theory)
    Theme
    Semantic Web
  9. Knowledge organization, information systems and other essays : professor A. Neelameghan Festschrift (2006) 0.11
    0.11380973 = product of:
      0.27314335 = sum of:
        0.014128185 = weight(_text_:web in 1487) [ClassicSimilarity], result of:
          0.014128185 = score(doc=1487,freq=2.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.12619963 = fieldWeight in 1487, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1487)
        0.014161124 = weight(_text_:information in 1487) [ClassicSimilarity], result of:
          0.014161124 = score(doc=1487,freq=24.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.23515764 = fieldWeight in 1487, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1487)
        0.14549215 = weight(_text_:wissensbasiertes in 1487) [ClassicSimilarity], result of:
          0.14549215 = score(doc=1487,freq=4.0), product of:
            0.30209768 = queryWeight, product of:
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.03430388 = queryNorm
            0.48160633 = fieldWeight in 1487, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1487)
        0.029423824 = weight(_text_:system in 1487) [ClassicSimilarity], result of:
          0.029423824 = score(doc=1487,freq=10.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.2723372 = fieldWeight in 1487, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1487)
        0.06993809 = product of:
          0.13987619 = sum of:
            0.13987619 = weight(_text_:aufsatzsammlung in 1487) [ClassicSimilarity], result of:
              0.13987619 = score(doc=1487,freq=12.0), product of:
                0.2250708 = queryWeight, product of:
                  6.5610886 = idf(docFreq=169, maxDocs=44218)
                  0.03430388 = queryNorm
                0.62147635 = fieldWeight in 1487, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  6.5610886 = idf(docFreq=169, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1487)
          0.5 = coord(1/2)
      0.41666666 = coord(5/12)
    
    Content
    Inhalt: KNOWLEDGE ORGANIZATION Towards a Future for Knowledge Organization Ingetraut Dahlberg Professor Neelameghan's Contribution to the Advancement and Development of Classification in the Context of Knowledge Organization Nancy J. Williamson Knowledge Orgnization System Over Time S. Seetharama The Facet Concept as a Universal Principle of Subdivisio Clare Beghtol Facet Analysis as a Knowledge Management Tool on the Internet Kathryn La Barre and Pauline Atherton Cochrane The Universal Decimal Classification: A Response to a Challenge I. C. Mellwaine Controlled Vocabularies as a Sphere of Influence Anita S. Coleman and Paul Bracke Aligning Systems of Relationship Rebecca Green and Carol A. Bean Terminologies, Ontologies and Information Access Widad Mustafa El Hadi SATSAN AUTOMATRIX Version 1 : A Computer Programme for Synthesis of Colon Class Number According to the Postulational Approach B. G. Satyapal and N. Sanjivini Satyapal. INTEROPERABILITY, DIGITAL LIBRARY AND INFORMATION RETRIEVAL Interoperable Institutional Digital Research Repositories and Their Potential for Open Access Research Knowledge Management T. B. Rajashekar Boundary Objects and the Digital Library Michael Shepherd and Corolyn Watters A PFT-based Approach to Make CDS/ISIS Data based OAI-Compliant Francis Jayakanth and L. Aswath The changing Language Technology and CDS/ ISIS: UNICODE and the Emergence of OTF K. H. Hussain and J. S. Rajeev Text Mining in Biomedicine: Challenges and Opportunities Padmini Srinivasan Determining Authorship of Web Pages Timothy C. Craven
    KNOWLEDGE MANAGEMENT IN SPECIALIZED AREAS Information System for Knowledge Management in the Specialized Division of a Hospital M. C. Vasudevan; Murali Mohan and Amit Kapoor Five Laws of Information Service and Architecting Knowledge Infrastructure for Education and Development k. R. Srivathsan Documentation of Compositions in Carnatic Music: Need for and Utility of a Computerized Database K. S. Nagarajan Saint Tyagaraja CD: A Model for Knowledge Organization and Presentation of Classical Carnatic Music---T. N. Rajan The National Tuberculosis Institute, Bangalore; Recent Development in Library and Information Services Sudha S. Murthy Sri Ramakrishna Math Libraries: Computer Applications D.N. Nagaraja Rao Save the Time of the Godly: Information Mediator's Role in Promoting Spiritual and Religious Accommodation Mohamed Taher INFORMATION SOCIETY Information Society, Information Networks and National Development : An Overview P. B. Mangla Digital Divide in India-Narrowing the Gap: An Appraisal with Special Reference to Karnataka K. N. Prasad Future of the Book: Will the Printed Book Survive the Digital Age? K. A. Isaac Role of Traditional Librarianship in the Internet/Digital Era a. Ratnakar A New Paradigm of Education System for Reaching the Unreached Through Open and Distance Education with Special Reference to the Indian Initiative S. B. Ghosh Knowledge Workers of the New Millennium: An Instance of Interdisciplinary Exchange and Discovery Michael Medland
    RSWK
    Wissensorganisation / Aufsatzsammlung (GBV)
    Indien / Wissensbasiertes System / Aufsatzsammlung (SWB)
    Indien / Informations- und Dokumentationswissenschaft / Aufsatzsammlung (SWB)
    Dewey-Dezimalklassifikation / Elektronische Bibliothek / Informationsgesellschaft / Information Retrieval (GBV)
    Subject
    Wissensorganisation / Aufsatzsammlung (GBV)
    Indien / Wissensbasiertes System / Aufsatzsammlung (SWB)
    Indien / Informations- und Dokumentationswissenschaft / Aufsatzsammlung (SWB)
    Dewey-Dezimalklassifikation / Elektronische Bibliothek / Informationsgesellschaft / Information Retrieval (GBV)
  10. Jung, H.; Yi, E.; Kim, D.; Lee, G.G.: Information extraction with automatic knowledge expansion (2005) 0.10
    0.10469659 = product of:
      0.31408975 = sum of:
        0.024219744 = weight(_text_:web in 1008) [ClassicSimilarity], result of:
          0.024219744 = score(doc=1008,freq=2.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.21634221 = fieldWeight in 1008, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=1008)
        0.017165873 = weight(_text_:information in 1008) [ClassicSimilarity], result of:
          0.017165873 = score(doc=1008,freq=12.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.2850541 = fieldWeight in 1008, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=1008)
        0.24080253 = weight(_text_:extraction in 1008) [ClassicSimilarity], result of:
          0.24080253 = score(doc=1008,freq=18.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            1.1815379 = fieldWeight in 1008, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.046875 = fieldNorm(doc=1008)
        0.031901587 = weight(_text_:system in 1008) [ClassicSimilarity], result of:
          0.031901587 = score(doc=1008,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.29527056 = fieldWeight in 1008, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.046875 = fieldNorm(doc=1008)
      0.33333334 = coord(4/12)
    
    Abstract
    POSIE (POSTECH Information Extraction System) is an information extraction system which uses multiple learning strategies, i.e., SmL, user-oriented learning, and separate-context learning, in a question answering framework. POSIE replaces laborious annotation with automatic instance extraction by the SmL from structured Web documents, and places the user at the end of the user-oriented learning cycle. Information extraction as question answering simplifies the extraction procedures for a set of slots. We introduce the techniques verified on the question answering framework, such as domain knowledge and instance rules, into an information extraction problem. To incrementally improve extraction performance, a sequence of the user-oriented learning and the separate-context learning produces context rules and generalizes them in both the learning and extraction phases. Experiments on the "continuing education" domain initially show that the F1-measure becomes 0.477 and recall 0.748 with no user training. However, as the size of the training documents grows, the F1-measure reaches beyond 0.75 with recall 0.772. We also obtain F-measure of about 0.9 for five out of seven slots on "job offering" domain.
    Source
    Information processing and management. 41(2005) no.2, S.217-242
  11. Heidorn, P.B.; Wei, Q.: Automatic metadata extraction from museum specimen labels (2008) 0.08
    0.08250257 = product of:
      0.19800617 = sum of:
        0.02018312 = weight(_text_:web in 2624) [ClassicSimilarity], result of:
          0.02018312 = score(doc=2624,freq=2.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.18028519 = fieldWeight in 2624, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2624)
        0.0058399485 = weight(_text_:information in 2624) [ClassicSimilarity], result of:
          0.0058399485 = score(doc=2624,freq=2.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.09697737 = fieldWeight in 2624, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2624)
        0.13377918 = weight(_text_:extraction in 2624) [ClassicSimilarity], result of:
          0.13377918 = score(doc=2624,freq=8.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.6564099 = fieldWeight in 2624, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2624)
        0.026584659 = weight(_text_:system in 2624) [ClassicSimilarity], result of:
          0.026584659 = score(doc=2624,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.24605882 = fieldWeight in 2624, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2624)
        0.011619256 = product of:
          0.023238512 = sum of:
            0.023238512 = weight(_text_:22 in 2624) [ClassicSimilarity], result of:
              0.023238512 = score(doc=2624,freq=2.0), product of:
                0.120126344 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03430388 = queryNorm
                0.19345059 = fieldWeight in 2624, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2624)
          0.5 = coord(1/2)
      0.41666666 = coord(5/12)
    
    Abstract
    This paper describes the information properties of museum specimen labels and machine learning tools to automatically extract Darwin Core (DwC) and other metadata from these labels processed through Optical Character Recognition (OCR). The DwC is a metadata profile describing the core set of access points for search and retrieval of natural history collections and observation databases. Using the HERBIS Learning System (HLS) we extract 74 independent elements from these labels. The automated text extraction tools are provided as a web service so that users can reference digital images of specimens and receive back an extended Darwin Core XML representation of the content of the label. This automated extraction task is made more difficult by the high variability of museum label formats, OCR errors and the open class nature of some elements. In this paper we introduce our overall system architecture, and variability robust solutions including, the application of Hidden Markov and Naïve Bayes machine learning models, data cleaning, use of field element identifiers, and specialist learning models. The techniques developed here could be adapted to any metadata extraction situation with noisy text and weakly ordered elements.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
  12. Hsu, C.-N.; Chang, C.-H.; Hsieh, C.-H.; Lu, J.-J.; Chang, C.-C.: Reconfigurable Web wrapper agents for biological information integration (2005) 0.06
    0.06386096 = product of:
      0.19158287 = sum of:
        0.06991638 = weight(_text_:web in 5263) [ClassicSimilarity], result of:
          0.06991638 = score(doc=5263,freq=24.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.6245262 = fieldWeight in 5263, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5263)
        0.015451051 = weight(_text_:information in 5263) [ClassicSimilarity], result of:
          0.015451051 = score(doc=5263,freq=14.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.256578 = fieldWeight in 5263, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5263)
        0.09459618 = weight(_text_:extraction in 5263) [ClassicSimilarity], result of:
          0.09459618 = score(doc=5263,freq=4.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.46415195 = fieldWeight in 5263, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5263)
        0.011619256 = product of:
          0.023238512 = sum of:
            0.023238512 = weight(_text_:22 in 5263) [ClassicSimilarity], result of:
              0.023238512 = score(doc=5263,freq=2.0), product of:
                0.120126344 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03430388 = queryNorm
                0.19345059 = fieldWeight in 5263, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5263)
          0.5 = coord(1/2)
      0.33333334 = coord(4/12)
    
    Abstract
    A variety of biological data is transferred and exchanged in overwhelming volumes on the World Wide Web. How to rapidly capture, utilize, and integrate the information on the Internet to discover valuable biological knowledge is one of the most critical issues in bioinformatics. Many information integration systems have been proposed for integrating biological data. These systems usually rely on an intermediate software layer called wrappers to access connected information sources. Wrapper construction for Web data sources is often specially hand coded to accommodate the differences between each Web site. However, programming a Web wrapper requires substantial programming skill, and is time-consuming and hard to maintain. In this article we provide a solution for rapidly building software agents that can serve as Web wrappers for biological information integration. We define an XML-based language called Web Navigation Description Language (WNDL), to model a Web-browsing session. A WNDL script describes how to locate the data, extract the data, and combine the data. By executing different WNDL scripts, we can automate virtually all types of Web-browsing sessions. We also describe IEPAD (Information Extraction Based on Pattern Discovery), a data extractor based on pattern discovery techniques. IEPAD allows our software agents to automatically discover the extraction rules to extract the contents of a structurally formatted Web page. With a programming-by-example authoring tool, a user can generate a complete Web wrapper agent by browsing the target Web sites. We built a variety of biological applications to demonstrate the feasibility of our approach.
    Date
    22. 7.2006 14:36:42
    Source
    Journal of the American Society for Information Science and Technology. 56(2005) no.5, S.505-517
  13. Radev, D.; Fan, W.; Qu, H.; Wu, H.; Grewal, A.: Probabilistic question answering on the Web (2005) 0.06
    0.061677005 = product of:
      0.18503101 = sum of:
        0.04194983 = weight(_text_:web in 3455) [ClassicSimilarity], result of:
          0.04194983 = score(doc=3455,freq=6.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.37471575 = fieldWeight in 3455, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=3455)
        0.0070079383 = weight(_text_:information in 3455) [ClassicSimilarity], result of:
          0.0070079383 = score(doc=3455,freq=2.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.116372846 = fieldWeight in 3455, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3455)
        0.113515414 = weight(_text_:extraction in 3455) [ClassicSimilarity], result of:
          0.113515414 = score(doc=3455,freq=4.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.55698234 = fieldWeight in 3455, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.046875 = fieldNorm(doc=3455)
        0.02255783 = weight(_text_:system in 3455) [ClassicSimilarity], result of:
          0.02255783 = score(doc=3455,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.20878783 = fieldWeight in 3455, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.046875 = fieldNorm(doc=3455)
      0.33333334 = coord(4/12)
    
    Abstract
    Web-based search engines such as Google and NorthernLight return documents that are relevant to a user query, not answers to user questions. We have developed an architecture that augments existing search engines so that they support natural language question answering. The process entails five steps: query modulation, document retrieval, passage extraction, phrase extraction, and answer ranking. In this article, we describe some probabilistic approaches to the last three of these stages. We show how our techniques apply to a number of existing search engines, and we also present results contrasting three different methods for question answering. Our algorithm, probabilistic phrase reranking (PPR), uses proximity and question type features and achieves a total reciprocal document rank of .20 an the TREC8 corpus. Our techniques have been implemented as a Web-accessible system, called NSIR.
    Source
    Journal of the American Society for Information Science and Technology. 56(2005) no.6, S.571-583
  14. Medien-Informationsmanagement : Archivarische, dokumentarische, betriebswirtschaftliche, rechtliche und Berufsbild-Aspekte ; [Frühjahrstagung der Fachgruppe 7 im Jahr 2000 in Weimar und Folgetagung 2001 in Köln] (2003) 0.06
    0.06092712 = product of:
      0.18278135 = sum of:
        0.012109872 = weight(_text_:web in 1833) [ClassicSimilarity], result of:
          0.012109872 = score(doc=1833,freq=2.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.108171105 = fieldWeight in 1833, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1833)
        0.0070079383 = weight(_text_:information in 1833) [ClassicSimilarity], result of:
          0.0070079383 = score(doc=1833,freq=8.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.116372846 = fieldWeight in 1833, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1833)
        0.011278915 = weight(_text_:system in 1833) [ClassicSimilarity], result of:
          0.011278915 = score(doc=1833,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.104393914 = fieldWeight in 1833, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1833)
        0.15238462 = sum of:
          0.13844152 = weight(_text_:aufsatzsammlung in 1833) [ClassicSimilarity], result of:
            0.13844152 = score(doc=1833,freq=16.0), product of:
              0.2250708 = queryWeight, product of:
                6.5610886 = idf(docFreq=169, maxDocs=44218)
                0.03430388 = queryNorm
              0.61510205 = fieldWeight in 1833, product of:
                4.0 = tf(freq=16.0), with freq of:
                  16.0 = termFreq=16.0
                6.5610886 = idf(docFreq=169, maxDocs=44218)
                0.0234375 = fieldNorm(doc=1833)
          0.013943106 = weight(_text_:22 in 1833) [ClassicSimilarity], result of:
            0.013943106 = score(doc=1833,freq=2.0), product of:
              0.120126344 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03430388 = queryNorm
              0.116070345 = fieldWeight in 1833, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0234375 = fieldNorm(doc=1833)
      0.33333334 = coord(4/12)
    
    Abstract
    Als in den siebziger Jahren des vergangenen Jahrhunderts immer häufiger die Bezeichnung Informationsmanager für Leute propagiert wurde, die bis dahin als Dokumentare firmierten, wurde dies in den etablierten Kreisen der Archivare und Bibliothekare gelegentlich belächelt und als Zeichen einer Identitätskrise oder jedenfalls einer Verunsicherung des damit überschriebenen Berufsbilds gewertet. Für den Berufsstand der Medienarchivare/Mediendokumentare, die sich seit 1960 in der Fachgruppe 7 des Vereins, später Verbands deutscher Archivare (VdA) organisieren, gehörte diese Verortung im Zeichen neuer inhaltlicher Herausforderungen (Informationsflut) und Technologien (EDV) allerdings schon früh zu den Selbstverständlichkeiten des Berufsalltags. "Halt, ohne uns geht es nicht!" lautete die Überschrift eines Artikels im Verbandsorgan "Info 7", der sich mit der Einrichtung von immer mächtigeren Leitungsnetzen und immer schnelleren Datenautobahnen beschäftigte. Information, Informationsgesellschaft: diese Begriffe wurden damals fast nur im technischen Sinne verstanden. Die informatisierte, nicht die informierte Gesellschaft stand im Vordergrund - was wiederum Kritiker auf den Plan rief, von Joseph Weizenbaum in den USA bis hin zu den Informations-Ökologen in Bremen. Bei den nationalen, manchmal auch nur regionalen Projekten und Modellversuchen mit Datenautobahnen - auch beim frühen Btx - war nie so recht deutlich geworden, welche Inhalte in welcher Gestalt durch diese Netze und Straßen gejagt werden sollten und wer diese Inhalte eigentlich selektieren, portionieren, positionieren, kurz: managen sollte. Spätestens mit dem World Wide Web sind diese Projekte denn auch obsolet geworden, jedenfalls was die Hardware und Software anging. Geblieben ist das Thema Inhalte (neudeutsch: Content). Und - immer drängender im nicht nur technischen Verständnis - das Thema Informationsmanagement. MedienInformationsManagement war die Frühjahrstagung der Fachgruppe 7 im Jahr 2000 in Weimar überschrieben, und auch die Folgetagung 2001 in Köln, die der multimedialen Produktion einen dokumentarischen Pragmatismus gegenüber stellte, handelte vom Geschäftsfeld Content und von Content-Management-Systemen. Die in diesem 6. Band der Reihe Beiträge zur Mediendokumentation versammelten Vorträge und Diskussionsbeiträge auf diesen beiden Tagungen beleuchten das Titel-Thema aus den verschiedensten Blickwinkeln: archivarischen, dokumentarischen, kaufmännischen, berufsständischen und juristischen. Deutlich wird dabei, daß die Berufsbezeichnung Medienarchivarln/Mediendokumentarln ziemlich genau für all das steht, was heute mit sog. alten wie neuen Medien im organisatorischen, d.h. ordnenden und vermittelnden Sinne geschieht. Im besonderen Maße trifft dies auf das Internet und die aus ihm geborenen Intranets zu. Beide bedürfen genauso der ordnenden Hand, die sich an den alten Medien, an Buch, Zeitung, Tonträger, Film etc. geschult hat, denn sie leben zu großen Teilen davon. Daß das Internet gleichwohl ein Medium sui generis ist und die alten Informationsberufe vor ganz neue Herausforderungen stellt - auch das durchzieht die Beiträge von Weimar und Köln.
    Content
    Enthält u.a. die Beiträge (Dokumentarische Aspekte): Günter Perers/Volker Gaese: Das DocCat-System in der Textdokumentation von Gr+J (Weimar 2000) Thomas Gerick: Finden statt suchen. Knowledge Retrieval in Wissensbanken. Mit organisiertem Wissen zu mehr Erfolg (Weimar 2000) Winfried Gödert: Aufbereitung und Rezeption von Information (Weimar 2000) Elisabeth Damen: Klassifikation als Ordnungssystem im elektronischen Pressearchiv (Köln 2001) Clemens Schlenkrich: Aspekte neuer Regelwerksarbeit - Multimediales Datenmodell für ARD und ZDF (Köln 2001) Josef Wandeler: Comprenez-vous only Bahnhof'? - Mehrsprachigkeit in der Mediendokumentation (Köln 200 1)
    Date
    11. 5.2008 19:49:22
    LCSH
    Information technology / Management / Congresses
    RSWK
    Mediendokumentation / Aufsatzsammlung
    Medien / Informationsmanagement / Aufsatzsammlung
    Pressearchiv / Aufsatzsammlung (HBZ)
    Rundfunkarchiv / Aufsatzsammlung (HBZ)
    Subject
    Mediendokumentation / Aufsatzsammlung
    Medien / Informationsmanagement / Aufsatzsammlung
    Pressearchiv / Aufsatzsammlung (HBZ)
    Rundfunkarchiv / Aufsatzsammlung (HBZ)
    Information technology / Management / Congresses
  15. Trkulja, V.: Suche ist überall, Semantic Web setzt sich durch, Renaissance der Taxonomien (2005) 0.06
    0.060767993 = product of:
      0.24307197 = sum of:
        0.06850378 = weight(_text_:web in 3295) [ClassicSimilarity], result of:
          0.06850378 = score(doc=3295,freq=4.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.6119082 = fieldWeight in 3295, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.09375 = fieldNorm(doc=3295)
        0.014015877 = weight(_text_:information in 3295) [ClassicSimilarity], result of:
          0.014015877 = score(doc=3295,freq=2.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.23274569 = fieldWeight in 3295, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=3295)
        0.16055231 = weight(_text_:suche in 3295) [ClassicSimilarity], result of:
          0.16055231 = score(doc=3295,freq=4.0), product of:
            0.17138755 = queryWeight, product of:
              4.996156 = idf(docFreq=812, maxDocs=44218)
              0.03430388 = queryNorm
            0.93677926 = fieldWeight in 3295, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.996156 = idf(docFreq=812, maxDocs=44218)
              0.09375 = fieldNorm(doc=3295)
      0.25 = coord(3/12)
    
    Abstract
    Ein Schwerpunkt der Online Information 2004 bildete das Thema "Search": Wie wird die Suche in 2005 aussehen? Welche Bedeutung haben Taxonomien? Wie verändern sich Suchfunktionen?
    Theme
    Semantic Web
  16. Naing, M.-M.; Lim, E.-P.; Chiang, R.H.L.: Extracting link chains of relationship instances from a Web site (2006) 0.06
    0.060050547 = product of:
      0.24020219 = sum of:
        0.07265923 = weight(_text_:web in 6111) [ClassicSimilarity], result of:
          0.07265923 = score(doc=6111,freq=18.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.64902663 = fieldWeight in 6111, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=6111)
        0.0070079383 = weight(_text_:information in 6111) [ClassicSimilarity], result of:
          0.0070079383 = score(doc=6111,freq=2.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.116372846 = fieldWeight in 6111, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=6111)
        0.16053502 = weight(_text_:extraction in 6111) [ClassicSimilarity], result of:
          0.16053502 = score(doc=6111,freq=8.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.78769195 = fieldWeight in 6111, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.046875 = fieldNorm(doc=6111)
      0.25 = coord(3/12)
    
    Abstract
    Web pages from a Web site can often be associated with concepts in an ontology, and pairs of Web pages also can be associated with relationships between concepts. With such associations, the Web site can be searched, browsed, or even reorganized based on the concept and relationship labels of its Web pages. In this article, we study the link chain extraction problem that is critical to the extraction of Web pages that are related. A link chain is an ordered list of anchor elements linking two Web pages related by some semantic relationship. We propose a link chain extraction method that derives extraction rules for identifying the anchor elements forming the link chains. We applied the proposed method to two well-structured Web sites and found that its performance in terms of precision and recall is good, even with a small number of training examples.
    Source
    Journal of the American Society for Information Science and Technology. 57(2006) no.12, S.1590-1605
  17. Zhou, G.D.; Zhang, M.: Extracting relation information from text documents by exploring various types of knowledge (2007) 0.06
    0.05501894 = product of:
      0.22007576 = sum of:
        0.016517868 = weight(_text_:information in 927) [ClassicSimilarity], result of:
          0.016517868 = score(doc=927,freq=16.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.27429342 = fieldWeight in 927, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=927)
        0.17697324 = weight(_text_:extraction in 927) [ClassicSimilarity], result of:
          0.17697324 = score(doc=927,freq=14.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.8683488 = fieldWeight in 927, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.0390625 = fieldNorm(doc=927)
        0.026584659 = weight(_text_:system in 927) [ClassicSimilarity], result of:
          0.026584659 = score(doc=927,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.24605882 = fieldWeight in 927, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=927)
      0.25 = coord(3/12)
    
    Abstract
    Extracting semantic relationships between entities from text documents is challenging in information extraction and important for deep information processing and management. This paper investigates the incorporation of diverse lexical, syntactic and semantic knowledge in feature-based relation extraction using support vector machines. Our study illustrates that the base phrase chunking information is very effective for relation extraction and contributes to most of the performance improvement from syntactic aspect while current commonly used features from full parsing give limited further enhancement. This suggests that most of useful information in full parse trees for relation extraction is shallow and can be captured by chunking. This indicates that a cheap and robust solution in relation extraction can be achieved without decreasing too much in performance. We also demonstrate how semantic information such as WordNet, can be used in feature-based relation extraction to further improve the performance. Evaluation on the ACE benchmark corpora shows that effective incorporation of diverse features enables our system outperform previously best-reported systems. It also shows that our feature-based system significantly outperforms tree kernel-based systems. This suggests that current tree kernels fail to effectively explore structured syntactic information in relation extraction.
    Source
    Information processing and management. 43(2007) no.4, S.969-982
  18. Hellweg, H.; Hermes, B.; Stempfhuber, M.; Enderle, W.; Fischer, T.: DBClear : a generic system for clearinghouses (2002) 0.05
    0.054030403 = product of:
      0.16209121 = sum of:
        0.03425189 = weight(_text_:web in 3605) [ClassicSimilarity], result of:
          0.03425189 = score(doc=3605,freq=4.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.3059541 = fieldWeight in 3605, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=3605)
        0.015670227 = weight(_text_:information in 3605) [ClassicSimilarity], result of:
          0.015670227 = score(doc=3605,freq=10.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.2602176 = fieldWeight in 3605, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3605)
        0.08026751 = weight(_text_:extraction in 3605) [ClassicSimilarity], result of:
          0.08026751 = score(doc=3605,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.39384598 = fieldWeight in 3605, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.046875 = fieldNorm(doc=3605)
        0.031901587 = weight(_text_:system in 3605) [ClassicSimilarity], result of:
          0.031901587 = score(doc=3605,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.29527056 = fieldWeight in 3605, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.046875 = fieldNorm(doc=3605)
      0.33333334 = coord(4/12)
    
    Abstract
    Clearinghouses - or subject gateways - are domain-specific collections of links to resources an the Internet. The links are described with metadata and structured according to a domain-specific subject hierarchy. Users access the information by searching in the metadata or by browsing the subject hierarchy. The standards for metadata vary across existing Clearinghouses and different technologies for storing and accessing the metadata are used. This makes it difficult to distribute the editorial or administrative work involved in maintaining a clearinghouse, or to exchange information with other systems. DBClear is a generic, platform-independent clearinghouse system, whose metadata schema can be adapted to different standards. The data is stored in a relational database. It includes a workflow component to Support distributed maintenance and automation modules for link checking and metadata extraction. The presentation of the clearinghouse an the Web can be modified to allow seamless integration into existing web sites.
    Source
    Gaining insight from research information (CRIS2002): Proceedings of the 6th International Conference an Current Research Information Systems, University of Kassel, August 29 - 31, 2002. Eds: W. Adamczak u. A. Nase
    Theme
    Information Gateway
  19. Schutz, A.; Buitelaar, P.: RelExt: a tool for relation extraction from text in ontology extension (2005) 0.05
    0.052661 = product of:
      0.157983 = sum of:
        0.028543243 = weight(_text_:web in 1078) [ClassicSimilarity], result of:
          0.028543243 = score(doc=1078,freq=4.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.25496176 = fieldWeight in 1078, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1078)
        0.008258934 = weight(_text_:information in 1078) [ClassicSimilarity], result of:
          0.008258934 = score(doc=1078,freq=4.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.13714671 = fieldWeight in 1078, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1078)
        0.09459618 = weight(_text_:extraction in 1078) [ClassicSimilarity], result of:
          0.09459618 = score(doc=1078,freq=4.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.46415195 = fieldWeight in 1078, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1078)
        0.026584659 = weight(_text_:system in 1078) [ClassicSimilarity], result of:
          0.026584659 = score(doc=1078,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.24605882 = fieldWeight in 1078, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1078)
      0.33333334 = coord(4/12)
    
    Abstract
    Domain ontologies very rarely model verbs as relations holding between concepts. However, the role of the verb as a central connecting element between concepts is undeniable. Verbs specify the interaction between the participants of some action or event by expressing relations between them. In parallel, it can be argued from an ontology engineering point of view that verbs express a relation between two classes that specify domain and range. The work described here is concerned with relation extraction for ontology extension along these lines. We describe a system (RelExt) that is capable of automatically identifying highly relevant triples (pairs of concepts connected by a relation) over concepts from an existing ontology. RelExt works by extracting relevant verbs and their grammatical arguments (i.e. terms) from a domain-specific text collection and computing corresponding relations through a combination of linguistic and statistical processing. The paper includes a detailed description of the system architecture and evaluation results on a constructed benchmark. RelExt has been developed in the context of the SmartWeb project, which aims at providing intelligent information services via mobile broadband devices on the FIFA World Cup that will be hosted in Germany in 2006. Such services include location based navigational information as well as question answering in the football domain.
    Source
    Semantic Web - ISWC 2005, 4th International Semantic Web Conference, ISWC 2005, Galway, Ireland, November 6-10, 2005, Proceedings. Eds.: Yolanda Gil, Enrico Motta, V. Richard Benjamins, Mark A. Musen
  20. Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany (2008) 0.05
    0.051642425 = product of:
      0.123941824 = sum of:
        0.02825637 = weight(_text_:web in 2668) [ClassicSimilarity], result of:
          0.02825637 = score(doc=2668,freq=8.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.25239927 = fieldWeight in 2668, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.02734375 = fieldNorm(doc=2668)
        0.008175928 = weight(_text_:information in 2668) [ClassicSimilarity], result of:
          0.008175928 = score(doc=2668,freq=8.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.13576832 = fieldWeight in 2668, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02734375 = fieldNorm(doc=2668)
        0.06621732 = weight(_text_:extraction in 2668) [ClassicSimilarity], result of:
          0.06621732 = score(doc=2668,freq=4.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.32490635 = fieldWeight in 2668, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.02734375 = fieldNorm(doc=2668)
        0.013158734 = weight(_text_:system in 2668) [ClassicSimilarity], result of:
          0.013158734 = score(doc=2668,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.1217929 = fieldWeight in 2668, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.02734375 = fieldNorm(doc=2668)
        0.008133478 = product of:
          0.016266957 = sum of:
            0.016266957 = weight(_text_:22 in 2668) [ClassicSimilarity], result of:
              0.016266957 = score(doc=2668,freq=2.0), product of:
                0.120126344 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03430388 = queryNorm
                0.1354154 = fieldWeight in 2668, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=2668)
          0.5 = coord(1/2)
      0.41666666 = coord(5/12)
    
    Abstract
    Metadata is a key aspect of our evolving infrastructure for information management, social computing, and scientific collaboration. DC-2008 will focus on metadata challenges, solutions, and innovation in initiatives and activities underlying semantic and social applications. Metadata is part of the fabric of social computing, which includes the use of wikis, blogs, and tagging for collaboration and participation. Metadata also underlies the development of semantic applications, and the Semantic Web - the representation and integration of multimedia knowledge structures on the basis of semantic models. These two trends flow together in applications such as Wikipedia, where authors collectively create structured information that can be extracted and used to enhance access to and use of information sources. Recent discussion has focused on how existing bibliographic standards can be expressed as Semantic Web vocabularies to facilitate the ingration of library and cultural heritage data with other types of data. Harnessing the efforts of content providers and end-users to link, tag, edit, and describe their information in interoperable ways ("participatory metadata") is a key step towards providing knowledge environments that are scalable, self-correcting, and evolvable. DC-2008 will explore conceptual and practical issues in the development and deployment of semantic and social applications to meet the needs of specific communities of practice.
    Content
    Carol Jean Godby, Devon Smith, Eric Childress: Encoding Application Profiles in a Computational Model of the Crosswalk. - Maria Elisabete Catarino, Ana Alice Baptista: Relating Folksonomies with Dublin Core. - Ed Summers, Antoine Isaac, Clay Redding, Dan Krech: LCSH, SKOS and Linked Data. - Xia Lin, Jiexun Li, Xiaohua Zhou: Theme Creation for Digital Collections. - Boris Lauser, Gudrun Johannsen, Caterina Caracciolo, Willem Robert van Hage, Johannes Keizer, Philipp Mayr: Comparing Human and Automatic Thesaurus Mapping Approaches in the Agricultural Domain. - P. Bryan Heidorn, Qin Wei: Automatic Metadata Extraction From Museum Specimen Labels. - Stuart Allen Sutton, Diny Golder: Achievement Standards Network (ASN): An Application Profile for Mapping K-12 Educational Resources to Achievement Standards. - Allen H. Renear, Karen M. Wickett, Richard J. Urban, David Dubin, Sarah L. Shreeves: Collection/Item Metadata Relationships. - Seth van Hooland, Yves Bontemps, Seth Kaufman: Answering the Call for more Accountability: Applying Data Profiling to Museum Metadata. - Thomas Margaritopoulos, Merkourios Margaritopoulos, Ioannis Mavridis, Athanasios Manitsaris: A Conceptual Framework for Metadata Quality Assessment. - Miao Chen, Xiaozhong Liu, Jian Qin: Semantic Relation Extraction from Socially-Generated Tags: A Methodology for Metadata Generation. - Hak Lae Kim, Simon Scerri, John G. Breslin, Stefan Decker, Hong Gee Kim: The State of the Art in Tag Ontologies: A Semantic Model for Tagging and Folksonomies. - Martin Malmsten: Making a Library Catalogue Part of the Semantic Web. - Philipp Mayr, Vivien Petras: Building a Terminology Network for Search: The KoMoHe Project. - Michael Panzer: Cool URIs for the DDC: Towards Web-scale Accessibility of a Large Classification System. - Barbara Levergood, Stefan Farrenkopf, Elisabeth Frasnelli: The Specification of the Language of the Field and Interoperability: Cross-language Access to Catalogues and Online Libraries (CACAO)

Authors

Languages

Types

  • a 6791
  • m 862
  • el 487
  • s 281
  • x 156
  • b 40
  • r 38
  • i 37
  • n 20
  • p 3
  • More… Less…

Themes

Subjects

Classifications