Search (1093 results, page 1 of 55)

  • × type_ss:"el"
  1. Kleineberg, M.: Context analysis and context indexing : formal pragmatics in knowledge organization (2014) 0.61
    0.61042273 = product of:
      1.2208455 = sum of:
        0.0908061 = product of:
          0.2724183 = sum of:
            0.2724183 = weight(_text_:3a in 1826) [ClassicSimilarity], result of:
              0.2724183 = score(doc=1826,freq=2.0), product of:
                0.29082868 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03430388 = queryNorm
                0.93669677 = fieldWeight in 1826, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1826)
          0.33333334 = coord(1/3)
        0.04036624 = weight(_text_:web in 1826) [ClassicSimilarity], result of:
          0.04036624 = score(doc=1826,freq=2.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.36057037 = fieldWeight in 1826, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.078125 = fieldNorm(doc=1826)
        0.2724183 = weight(_text_:2f in 1826) [ClassicSimilarity], result of:
          0.2724183 = score(doc=1826,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.93669677 = fieldWeight in 1826, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.078125 = fieldNorm(doc=1826)
        0.2724183 = weight(_text_:2f in 1826) [ClassicSimilarity], result of:
          0.2724183 = score(doc=1826,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.93669677 = fieldWeight in 1826, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.078125 = fieldNorm(doc=1826)
        0.2724183 = weight(_text_:2f in 1826) [ClassicSimilarity], result of:
          0.2724183 = score(doc=1826,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.93669677 = fieldWeight in 1826, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.078125 = fieldNorm(doc=1826)
        0.2724183 = weight(_text_:2f in 1826) [ClassicSimilarity], result of:
          0.2724183 = score(doc=1826,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.93669677 = fieldWeight in 1826, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.078125 = fieldNorm(doc=1826)
      0.5 = coord(6/12)
    
    Source
    http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=5&ved=0CDQQFjAE&url=http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F3131107&ei=HzFWVYvGMsiNsgGTyoFI&usg=AFQjCNE2FHUeR9oQTQlNC4TPedv4Mo3DaQ&sig2=Rlzpr7a3BLZZkqZCXXN_IA&bvm=bv.93564037,d.bGg&cad=rja
  2. Popper, K.R.: Three worlds : the Tanner lecture on human values. Deliverd at the University of Michigan, April 7, 1978 (1978) 0.39
    0.39349306 = product of:
      0.9443833 = sum of:
        0.072644874 = product of:
          0.21793462 = sum of:
            0.21793462 = weight(_text_:3a in 230) [ClassicSimilarity], result of:
              0.21793462 = score(doc=230,freq=2.0), product of:
                0.29082868 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03430388 = queryNorm
                0.7493574 = fieldWeight in 230, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0625 = fieldNorm(doc=230)
          0.33333334 = coord(1/3)
        0.21793462 = weight(_text_:2f in 230) [ClassicSimilarity], result of:
          0.21793462 = score(doc=230,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.7493574 = fieldWeight in 230, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0625 = fieldNorm(doc=230)
        0.21793462 = weight(_text_:2f in 230) [ClassicSimilarity], result of:
          0.21793462 = score(doc=230,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.7493574 = fieldWeight in 230, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0625 = fieldNorm(doc=230)
        0.21793462 = weight(_text_:2f in 230) [ClassicSimilarity], result of:
          0.21793462 = score(doc=230,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.7493574 = fieldWeight in 230, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0625 = fieldNorm(doc=230)
        0.21793462 = weight(_text_:2f in 230) [ClassicSimilarity], result of:
          0.21793462 = score(doc=230,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.7493574 = fieldWeight in 230, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0625 = fieldNorm(doc=230)
      0.41666666 = coord(5/12)
    
    Source
    https%3A%2F%2Ftannerlectures.utah.edu%2F_documents%2Fa-to-z%2Fp%2Fpopper80.pdf&usg=AOvVaw3f4QRTEH-OEBmoYr2J_c7H
  3. Shala, E.: ¬Die Autonomie des Menschen und der Maschine : gegenwärtige Definitionen von Autonomie zwischen philosophischem Hintergrund und technologischer Umsetzbarkeit (2014) 0.31
    0.30521137 = product of:
      0.61042273 = sum of:
        0.04540305 = product of:
          0.13620915 = sum of:
            0.13620915 = weight(_text_:3a in 4388) [ClassicSimilarity], result of:
              0.13620915 = score(doc=4388,freq=2.0), product of:
                0.29082868 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03430388 = queryNorm
                0.46834838 = fieldWeight in 4388, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4388)
          0.33333334 = coord(1/3)
        0.02018312 = weight(_text_:web in 4388) [ClassicSimilarity], result of:
          0.02018312 = score(doc=4388,freq=2.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.18028519 = fieldWeight in 4388, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4388)
        0.13620915 = weight(_text_:2f in 4388) [ClassicSimilarity], result of:
          0.13620915 = score(doc=4388,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.46834838 = fieldWeight in 4388, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4388)
        0.13620915 = weight(_text_:2f in 4388) [ClassicSimilarity], result of:
          0.13620915 = score(doc=4388,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.46834838 = fieldWeight in 4388, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4388)
        0.13620915 = weight(_text_:2f in 4388) [ClassicSimilarity], result of:
          0.13620915 = score(doc=4388,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.46834838 = fieldWeight in 4388, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4388)
        0.13620915 = weight(_text_:2f in 4388) [ClassicSimilarity], result of:
          0.13620915 = score(doc=4388,freq=2.0), product of:
            0.29082868 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03430388 = queryNorm
            0.46834838 = fieldWeight in 4388, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4388)
      0.5 = coord(6/12)
    
    Footnote
    Vgl. unter: https://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=2&cad=rja&uact=8&ved=2ahUKEwizweHljdbcAhVS16QKHXcFD9QQFjABegQICRAB&url=https%3A%2F%2Fwww.researchgate.net%2Fpublication%2F271200105_Die_Autonomie_des_Menschen_und_der_Maschine_-_gegenwartige_Definitionen_von_Autonomie_zwischen_philosophischem_Hintergrund_und_technologischer_Umsetzbarkeit_Redigierte_Version_der_Magisterarbeit_Karls&usg=AOvVaw06orrdJmFF2xbCCp_hL26q.
  4. Kara, S.: ¬An ontology-based retrieval system using semantic indexing (2012) 0.07
    0.07412624 = product of:
      0.22237872 = sum of:
        0.024219744 = weight(_text_:web in 3829) [ClassicSimilarity], result of:
          0.024219744 = score(doc=3829,freq=2.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.21634221 = fieldWeight in 3829, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=3829)
        0.014015877 = weight(_text_:information in 3829) [ClassicSimilarity], result of:
          0.014015877 = score(doc=3829,freq=8.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.23274569 = fieldWeight in 3829, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3829)
        0.13902743 = weight(_text_:extraction in 3829) [ClassicSimilarity], result of:
          0.13902743 = score(doc=3829,freq=6.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.68216133 = fieldWeight in 3829, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.046875 = fieldNorm(doc=3829)
        0.04511566 = weight(_text_:system in 3829) [ClassicSimilarity], result of:
          0.04511566 = score(doc=3829,freq=8.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.41757566 = fieldWeight in 3829, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.046875 = fieldNorm(doc=3829)
      0.33333334 = coord(4/12)
    
    Abstract
    In this thesis, we present an ontology-based information extraction and retrieval system and its application to soccer domain. In general, we deal with three issues in semantic search, namely, usability, scalability and retrieval performance. We propose a keyword-based semantic retrieval approach. The performance of the system is improved considerably using domain-specific information extraction, inference and rules. Scalability is achieved by adapting a semantic indexing approach. The system is implemented using the state-of-the-art technologies in SemanticWeb and its performance is evaluated against traditional systems as well as the query expansion methods. Furthermore, a detailed evaluation is provided to observe the performance gain due to domain-specific information extraction and inference. Finally, we show how we use semantic indexing to solve simple structural ambiguities.
    Source
    Information Systems. 37(2012) no. 4, S.294-305
    Theme
    Semantic Web
  5. Rajasurya, S.; Muralidharan, T.; Devi, S.; Swamynathan, S.: Semantic information retrieval using ontology in university domain (2012) 0.05
    0.047985192 = product of:
      0.14395557 = sum of:
        0.04036624 = weight(_text_:web in 2861) [ClassicSimilarity], result of:
          0.04036624 = score(doc=2861,freq=8.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.36057037 = fieldWeight in 2861, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2861)
        0.010115089 = weight(_text_:information in 2861) [ClassicSimilarity], result of:
          0.010115089 = score(doc=2861,freq=6.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.16796975 = fieldWeight in 2861, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2861)
        0.06688959 = weight(_text_:extraction in 2861) [ClassicSimilarity], result of:
          0.06688959 = score(doc=2861,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.32820496 = fieldWeight in 2861, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2861)
        0.026584659 = weight(_text_:system in 2861) [ClassicSimilarity], result of:
          0.026584659 = score(doc=2861,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.24605882 = fieldWeight in 2861, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2861)
      0.33333334 = coord(4/12)
    
    Abstract
    Today's conventional search engines hardly do provide the essential content relevant to the user's search query. This is because the context and semantics of the request made by the user is not analyzed to the full extent. So here the need for a semantic web search arises. SWS is upcoming in the area of web search which combines Natural Language Processing and Artificial Intelligence. The objective of the work done here is to design, develop and implement a semantic search engine- SIEU(Semantic Information Extraction in University Domain) confined to the university domain. SIEU uses ontology as a knowledge base for the information retrieval process. It is not just a mere keyword search. It is one layer above what Google or any other search engines retrieve by analyzing just the keywords. Here the query is analyzed both syntactically and semantically. The developed system retrieves the web results more relevant to the user query through keyword expansion. The results obtained here will be accurate enough to satisfy the request made by the user. The level of accuracy will be enhanced since the query is analyzed semantically. The system will be of great use to the developers and researchers who work on web. The Google results are re-ranked and optimized for providing the relevant links. For ranking an algorithm has been applied which fetches more apt results for the user query.
  6. Dietze, S.; Maynard, D.; Demidova, E.; Risse, T.; Stavrakas, Y.: Entity extraction and consolidation for social Web content preservation (2012) 0.05
    0.046166122 = product of:
      0.18466449 = sum of:
        0.06054936 = weight(_text_:web in 470) [ClassicSimilarity], result of:
          0.06054936 = score(doc=470,freq=18.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.5408555 = fieldWeight in 470, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=470)
        0.008258934 = weight(_text_:information in 470) [ClassicSimilarity], result of:
          0.008258934 = score(doc=470,freq=4.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.13714671 = fieldWeight in 470, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=470)
        0.115856186 = weight(_text_:extraction in 470) [ClassicSimilarity], result of:
          0.115856186 = score(doc=470,freq=6.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.56846774 = fieldWeight in 470, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.0390625 = fieldNorm(doc=470)
      0.25 = coord(3/12)
    
    Abstract
    With the rapidly increasing pace at which Web content is evolving, particularly social media, preserving the Web and its evolution over time becomes an important challenge. Meaningful analysis of Web content lends itself to an entity-centric view to organise Web resources according to the information objects related to them. Therefore, the crucial challenge is to extract, detect and correlate entities from a vast number of heterogeneous Web resources where the nature and quality of the content may vary heavily. While a wealth of information extraction tools aid this process, we believe that, the consolidation of automatically extracted data has to be treated as an equally important step in order to ensure high quality and non-ambiguity of generated data. In this paper we present an approach which is based on an iterative cycle exploiting Web data for (1) targeted archiving/crawling of Web objects, (2) entity extraction, and detection, and (3) entity correlation. The long-term goal is to preserve Web content over time and allow its navigation and analysis based on well-formed structured RDF data about entities.
  7. García Barrios, V.M.: Informationsaufbereitung und Wissensorganisation in transnationalen Konzernen : Konzeption eines Informationssystems für große und geographisch verteilte Unternehmen mit dem Hyperwave Information System (2002) 0.04
    0.043970633 = product of:
      0.17588253 = sum of:
        0.010115089 = weight(_text_:information in 6000) [ClassicSimilarity], result of:
          0.010115089 = score(doc=6000,freq=6.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.16796975 = fieldWeight in 6000, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6000)
        0.14696926 = weight(_text_:wissensbasiertes in 6000) [ClassicSimilarity], result of:
          0.14696926 = score(doc=6000,freq=2.0), product of:
            0.30209768 = queryWeight, product of:
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.03430388 = queryNorm
            0.48649582 = fieldWeight in 6000, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.806516 = idf(docFreq=17, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6000)
        0.018798191 = weight(_text_:system in 6000) [ClassicSimilarity], result of:
          0.018798191 = score(doc=6000,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.17398985 = fieldWeight in 6000, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6000)
      0.25 = coord(3/12)
    
    Abstract
    Transnationale Konzerne haben einen dringenden Bedarf an einer umfassenden Lösung für ihre Intranetsysteme. Die spezifischen Anforderungen an ein wissensbasiertes Informationssystem sind vielfältig, die kritischsten davon sind aber allgemein gültig und ergeben sich aufgrund der stark vernetzten und geographisch verteilten Struktur des Konzerns. In unterschiedlichen Wissensdisziplinen, insbesondere Wissensmanagement, Informationsmanagement, Datenmanagement und Wissensorganisation, versucht man, oftmals in den einzelnen Disziplinen isoliert, die Erfüllung spezifischer Anforderungen zu implementieren. Nicht selten auf eine ineffektive Weise. Die nachfolgende Arbeit verfolgt deshalb einen ganzheitlichen Ansatz über die Wissensdisziplinen, um den umfangreichen Anforderungen gerecht werden zu können. Im Untersuchungsbereich der vorliegenden Arbeit wird die Problematik aus der Sicht der wichtigsten involvierten Wissensdisziplinen beleuchtet, um nach vorhandenen bzw. etablierten Lösungsansätzen zu suchen. Die spezifischen Einflussbereiche der Disziplinen auf Intranetlösungen werden überprüft und kritischen Aspekten von Anforderungen (beispielsweise 'starke örtliche Verteilung vs. Systemtransparenz', 'Replikationsmaßnahmen vs. Systemperformanz' oder 'semantische Wissensmodelle vs. bedarfsgerechten Wissenszugang') gegenübergestellt. Jede Disziplin bietet effiziente und effektive Lösungen für unterschiedliche Aspekte, es konnte jedoch kein umfassendes Gestaltungsmodell, welches die spezifischen Lösungsansätze der Disziplinen vereint, im Rahmen des Rechercheprozesses identifiziert werden. Aufgrund des oben beschriebenen Sachverhalts wird im Gestaltungsbereich dieser Arbeit ein zweiteiliges Technisches Gestaltungsmodell vorgestellt. Es besteht aus einem strategischen Analyseschema und einem funktionalen Komponentenschema, und berücksichtigt die Einflussbereiche oben erwähnter Wissensdisziplinen. Basierend auf der konkreten Anforderung einer Intranetlösung für einen transnationalen - und anonymisiert dargestellten - Konzern, wird das vorgestellte Modell angewandt, und auf Basis des Hyperwave Information Servers die technische Umsetzung eines wissensbasierten Informationssystems, von dem beispielhaft zwei Module näher beschrieben werden, gezeigt.
    Theme
    Information Resources Management
  8. Auer, S.; Bizer, C.; Kobilarov, G.; Lehmann, J.; Cyganiak, R.; Ives, Z.: DBpedia: a nucleus for a Web of open data (2007) 0.04
    0.042248793 = product of:
      0.16899517 = sum of:
        0.076589555 = weight(_text_:web in 4260) [ClassicSimilarity], result of:
          0.076589555 = score(doc=4260,freq=20.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.6841342 = fieldWeight in 4260, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4260)
        0.012138106 = weight(_text_:information in 4260) [ClassicSimilarity], result of:
          0.012138106 = score(doc=4260,freq=6.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.20156369 = fieldWeight in 4260, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=4260)
        0.08026751 = weight(_text_:extraction in 4260) [ClassicSimilarity], result of:
          0.08026751 = score(doc=4260,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.39384598 = fieldWeight in 4260, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.046875 = fieldNorm(doc=4260)
      0.25 = coord(3/12)
    
    Abstract
    DBpedia is a community effort to extract structured information from Wikipedia and to make this information available on the Web. DBpedia allows you to ask sophisticated queries against datasets derived from Wikipedia and to link other datasets on the Web to Wikipedia data. We describe the extraction of the DBpedia datasets, and how the resulting information is published on the Web for human- and machineconsumption. We describe some emerging applications from the DBpedia community and show how website authors can facilitate DBpedia content within their sites. Finally, we present the current status of interlinking DBpedia with other open datasets on the Web and outline how DBpedia could serve as a nucleus for an emerging Web of open data.
    Source
    ¬The Semantic Web : 6th International Semantic Web Conference, 2nd Asian Semantic Web Conference, ISWC 2007 + ASWC 2007, Busan, Korea, November 11-15, 2007 : proceedings. Ed.: Karl Aberer et al
    Theme
    Semantic Web
  9. Semantische Suche über 500 Millionen Web-Dokumente (2009) 0.04
    0.037818365 = product of:
      0.15127346 = sum of:
        0.048439488 = weight(_text_:web in 2434) [ClassicSimilarity], result of:
          0.048439488 = score(doc=2434,freq=8.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.43268442 = fieldWeight in 2434, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=2434)
        0.080276154 = weight(_text_:suche in 2434) [ClassicSimilarity], result of:
          0.080276154 = score(doc=2434,freq=4.0), product of:
            0.17138755 = queryWeight, product of:
              4.996156 = idf(docFreq=812, maxDocs=44218)
              0.03430388 = queryNorm
            0.46838963 = fieldWeight in 2434, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.996156 = idf(docFreq=812, maxDocs=44218)
              0.046875 = fieldNorm(doc=2434)
        0.02255783 = weight(_text_:system in 2434) [ClassicSimilarity], result of:
          0.02255783 = score(doc=2434,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.20878783 = fieldWeight in 2434, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.046875 = fieldNorm(doc=2434)
      0.25 = coord(3/12)
    
    Content
    "Wissenschaftler an der University of Washington haben eine neue Suchmaschinen-Engine geschrieben, die Zusammenhänge und Fakten aus mehr als 500 Millionen einzelner Web-Seiten zusammentragen kann. Das Werkzeug extrahiert dabei Informationen aus Milliarden von Textzeilen, indem die grundlegenden sprachlichen Beziehungen zwischen Wörtern analysiert werden. Experten glauben, dass solche Systeme zur automatischen Informationsgewinnung eines Tages die Grundlage deutlich smarterer Suchmaschinen bilden werden, als sie heute verfügbar sind. Dazu werden die wichtigsten Datenhappen zunächst von einem Algorithmus intern begutachtet und dann intelligent kombiniert, berichtet Technology Review in seiner Online-Ausgabe. Das Projekt US-Forscher stellt eine deutliche Ausweitung einer zuvor an der gleichen Hochschule entwickelten Technik namens TextRunner dar. Sowohl die Anzahl analysierbarer Seiten als auch die Themengebiete wurden dabei stark erweitert. "TextRunner ist deshalb so bedeutsam, weil es skaliert, ohne dass dabei ein Mensch eingreifen müsste", sagt Peter Norvig, Forschungsdirektor bei Google. Der Internet-Konzern spendete dem Projekt die riesige Datenbank aus einzelnen Web-Seiten, die TextRunner analysiert. "Das System kann Millionen von Beziehungen erkennen und erlernen - und zwar nicht nur jede einzeln. Einen Betreuer braucht die Software nicht, die Informationen werden selbstständig ermittelt.""
    Source
    http://www.heise.de/newsticker/Semantische-Suche-ueber-500-Millionen-Web-Dokumente--/meldung/140630
  10. Birmingham, W.; Pardo, B.; Meek, C.; Shifrin, J.: ¬The MusArt music-retrieval system (2002) 0.04
    0.03767684 = product of:
      0.11303052 = sum of:
        0.022834593 = weight(_text_:web in 1205) [ClassicSimilarity], result of:
          0.022834593 = score(doc=1205,freq=4.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.2039694 = fieldWeight in 1205, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=1205)
        0.0066071474 = weight(_text_:information in 1205) [ClassicSimilarity], result of:
          0.0066071474 = score(doc=1205,freq=4.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.10971737 = fieldWeight in 1205, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=1205)
        0.053511675 = weight(_text_:extraction in 1205) [ClassicSimilarity], result of:
          0.053511675 = score(doc=1205,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.26256397 = fieldWeight in 1205, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03125 = fieldNorm(doc=1205)
        0.030077105 = weight(_text_:system in 1205) [ClassicSimilarity], result of:
          0.030077105 = score(doc=1205,freq=8.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.27838376 = fieldWeight in 1205, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03125 = fieldNorm(doc=1205)
      0.33333334 = coord(4/12)
    
    Abstract
    Music websites are ubiquitous, and music downloads, such as MP3, are a major source of Web traffic. As the amount of musical content increases and the Web becomes an important mechanism for distributing music, we expect to see a rising demand for music search services. Many currently available music search engines rely on file names, song title, composer or performer as the indexing and retrieval mechanism. These systems do not make use of the musical content. We believe that a more natural, effective, and usable music-information retrieval (MIR) system should have audio input, where the user can query with musical content. We are developing a system called MusArt for audio-input MIR. With MusArt, as with other audio-input MIR systems, a user sings or plays a theme, hook, or riff from the desired piece of music. The system transcribes the query and searches for related themes in a database, returning the most similar themes, given some measure of similarity. We call this "retrieval by query." In this paper, we describe the architecture of MusArt. An important element of MusArt is metadata creation: we believe that it is essential to automatically abstract important musical elements, particularly themes. Theme extraction is performed by a subsystem called MME, which we describe later in this paper. Another important element of MusArt is its support for a variety of search engines, as we believe that MIR is too complex for a single approach to work for all queries. Currently, MusArt supports a dynamic time-warping search engine that has high recall, and a complementary stochastic search engine that searches over themes, emphasizing speed and relevancy. The stochastic search engine is discussed in this paper.
    Theme
    Information Gateway
  11. Dias, G.: Multiword unit hybrid extraction (o.J.) 0.04
    0.036851138 = product of:
      0.14740455 = sum of:
        0.008175928 = weight(_text_:information in 643) [ClassicSimilarity], result of:
          0.008175928 = score(doc=643,freq=2.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.13576832 = fieldWeight in 643, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=643)
        0.09364543 = weight(_text_:extraction in 643) [ClassicSimilarity], result of:
          0.09364543 = score(doc=643,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.45948696 = fieldWeight in 643, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.0546875 = fieldNorm(doc=643)
        0.045583192 = weight(_text_:system in 643) [ClassicSimilarity], result of:
          0.045583192 = score(doc=643,freq=6.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.42190298 = fieldWeight in 643, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0546875 = fieldNorm(doc=643)
      0.25 = coord(3/12)
    
    Abstract
    This paper describes an original hybrid system that extracts multiword unit candidates from part-of-speech tagged corpora. While classical hybrid systems manually define local part-of-speech patterns that lead to the identification of well-known multiword units (mainly compound nouns), our solution automatically identifies relevant syntactical patterns from the corpus. Word statistics are then combined with the endogenously acquired linguistic information in order to extract the most relevant sequences of words. As a result, (1) human intervention is avoided providing total flexibility of use of the system and (2) different multiword units like phrasal verbs, adverbial locutions and prepositional locutions may be identified. The system has been tested on the Brown Corpus leading to encouraging results
  12. Urs, S.R.; Angrosh, M.A.: Ontology-based knowledge organization systems in digital libraries : a comparison of experiments in OWL and KAON ontologies (2006 (?)) 0.04
    0.03684422 = product of:
      0.11053266 = sum of:
        0.027966553 = weight(_text_:web in 2799) [ClassicSimilarity], result of:
          0.027966553 = score(doc=2799,freq=6.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.24981049 = fieldWeight in 2799, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=2799)
        0.014015876 = weight(_text_:information in 2799) [ClassicSimilarity], result of:
          0.014015876 = score(doc=2799,freq=18.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.23274568 = fieldWeight in 2799, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=2799)
        0.053511675 = weight(_text_:extraction in 2799) [ClassicSimilarity], result of:
          0.053511675 = score(doc=2799,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.26256397 = fieldWeight in 2799, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03125 = fieldNorm(doc=2799)
        0.015038553 = weight(_text_:system in 2799) [ClassicSimilarity], result of:
          0.015038553 = score(doc=2799,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.13919188 = fieldWeight in 2799, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03125 = fieldNorm(doc=2799)
      0.33333334 = coord(4/12)
    
    Abstract
    Grounded on a strong belief that ontologies enhance the performance of information retrieval systems, there has been an upsurge of interest in ontologies. Its importance is identified in diverse research fields such as knowledge engineering, knowledge representation, qualitative modeling, language engineering, database design, information integration, object-oriented analysis, information retrieval and extraction, knowledge management and agent-based systems design (Guarino, 1998). While the role-played by ontologies, automatically lends a place of legitimacy for these tools, research in this area gains greater significance in the wake of various challenges faced in the contemporary digital environment. With the objective of overcoming various pitfalls associated with current search mechanisms, ontologies are increasingly used for developing efficient information retrieval systems. An indicator of research interest in the area of ontology is the Swoogle, a search engine for Semantic Web documents, terms and data found on the Web (Ding, Li et al, 2004). Given the complex nature of the digital content archived in digital libraries, ontologies can be employed for designing efficient forms of information retrieval in digital libraries. Knowledge representation assumes greater significance due to its crucial role in ontology development. These systems aid in developing intelligent information systems, wherein the notion of intelligence implies the ability of the system to find implicit consequences of its explicitly represented knowledge (Baader and Nutt, 2003). Knowledge representation formalisms such as 'Description Logics' are used to obtain explicit knowledge representation of the subject domain. These representations are developed into ontologies, which are used for developing intelligent information systems. Against this backdrop, the paper examines the use of Description Logics for conceptually modeling a chosen domain, which would be utilized for developing domain ontologies. The knowledge representation languages identified for this purpose are Web Ontology Language (OWL) and KArlsruhe ONtology (KAON) language. Drawing upon the various technical constructs in developing ontology-based information systems, the paper explains the working of the prototypes and also presents a comparative study of the two prototypes.
    Theme
    Information Gateway
  13. Klic, L.; Miller, M.; Nelson, J.K.; Germann, J.E.: Approaching the largest 'API' : extracting information from the Internet with Python (2018) 0.04
    0.035211276 = product of:
      0.1408451 = sum of:
        0.048439488 = weight(_text_:web in 4239) [ClassicSimilarity], result of:
          0.048439488 = score(doc=4239,freq=8.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.43268442 = fieldWeight in 4239, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4239)
        0.012138106 = weight(_text_:information in 4239) [ClassicSimilarity], result of:
          0.012138106 = score(doc=4239,freq=6.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.20156369 = fieldWeight in 4239, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=4239)
        0.08026751 = weight(_text_:extraction in 4239) [ClassicSimilarity], result of:
          0.08026751 = score(doc=4239,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.39384598 = fieldWeight in 4239, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.046875 = fieldNorm(doc=4239)
      0.25 = coord(3/12)
    
    Abstract
    This article explores the need for libraries to algorithmically access and manipulate the world's largest API: the Internet. The billions of pages on the 'Internet API' (HTTP, HTML, CSS, XPath, DOM, etc.) are easily accessible and manipulable. Libraries can assist in creating meaning through the datafication of information on the world wide web. Because most information is created for human consumption, some programming is required for automated extraction. Python is an easy-to-learn programming language with extensive packages and community support for web page automation. Four packages (Urllib, Selenium, BeautifulSoup, Scrapy) in Python can automate almost any web page for all sized projects. An example warrant data project is explained to illustrate how well Python packages can manipulate web pages to create meaning through assembling custom datasets.
  14. Knorz, G.; Rein, B.: Semantische Suche in einer Hochschulontologie : Ontologie-basiertes Information-Filtering und -Retrieval mit relationalen Datenbanken (2005) 0.03
    0.03478675 = product of:
      0.139147 = sum of:
        0.008175928 = weight(_text_:information in 4324) [ClassicSimilarity], result of:
          0.008175928 = score(doc=4324,freq=2.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.13576832 = fieldWeight in 4324, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4324)
        0.11470412 = weight(_text_:suche in 4324) [ClassicSimilarity], result of:
          0.11470412 = score(doc=4324,freq=6.0), product of:
            0.17138755 = queryWeight, product of:
              4.996156 = idf(docFreq=812, maxDocs=44218)
              0.03430388 = queryNorm
            0.6692675 = fieldWeight in 4324, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.996156 = idf(docFreq=812, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4324)
        0.016266957 = product of:
          0.032533914 = sum of:
            0.032533914 = weight(_text_:22 in 4324) [ClassicSimilarity], result of:
              0.032533914 = score(doc=4324,freq=2.0), product of:
                0.120126344 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03430388 = queryNorm
                0.2708308 = fieldWeight in 4324, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4324)
          0.5 = coord(1/2)
      0.25 = coord(3/12)
    
    Abstract
    Ontologien werden eingesetzt, um durch semantische Fundierung insbesondere für das Dokumentenretrieval eine grundlegend bessere Basis zu haben, als dies gegenwärtiger Stand der Technik ist. Vorgestellt wird eine an der FH Darmstadt entwickelte und eingesetzte Ontologie, die den Gegenstandsbereich Hochschule sowohl breit abdecken und gleichzeitig differenziert semantisch beschreiben soll. Das Problem der semantischen Suche besteht nun darin, dass sie für Informationssuchende so einfach wie bei gängigen Suchmaschinen zu nutzen sein soll, und gleichzeitig auf der Grundlage des aufwendigen Informationsmodells hochwertige Ergebnisse liefern muss. Es wird beschrieben, welche Möglichkeiten die verwendete Software K-Infinity bereitstellt und mit welchem Konzept diese Möglichkeiten für eine semantische Suche nach Dokumenten und anderen Informationseinheiten (Personen, Veranstaltungen, Projekte etc.) eingesetzt werden.
    Date
    11. 2.2011 18:22:25
  15. Eckert, K.: SKOS: eine Sprache für die Übertragung von Thesauri ins Semantic Web (2011) 0.03
    0.033333898 = product of:
      0.13333559 = sum of:
        0.07220933 = weight(_text_:web in 4331) [ClassicSimilarity], result of:
          0.07220933 = score(doc=4331,freq=10.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.6450079 = fieldWeight in 4331, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0625 = fieldNorm(doc=4331)
        0.04253545 = weight(_text_:system in 4331) [ClassicSimilarity], result of:
          0.04253545 = score(doc=4331,freq=4.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.3936941 = fieldWeight in 4331, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0625 = fieldNorm(doc=4331)
        0.01859081 = product of:
          0.03718162 = sum of:
            0.03718162 = weight(_text_:22 in 4331) [ClassicSimilarity], result of:
              0.03718162 = score(doc=4331,freq=2.0), product of:
                0.120126344 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03430388 = queryNorm
                0.30952093 = fieldWeight in 4331, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4331)
          0.5 = coord(1/2)
      0.25 = coord(3/12)
    
    Abstract
    Das Semantic Web - bzw. Linked Data - hat das Potenzial, die Verfügbarkeit von Daten und Wissen, sowie den Zugriff darauf zu revolutionieren. Einen großen Beitrag dazu können Wissensorganisationssysteme wie Thesauri leisten, die die Daten inhaltlich erschließen und strukturieren. Leider sind immer noch viele dieser Systeme lediglich in Buchform oder in speziellen Anwendungen verfügbar. Wie also lassen sie sich für das Semantic Web nutzen? Das Simple Knowledge Organization System (SKOS) bietet eine Möglichkeit, die Wissensorganisationssysteme in eine Form zu "übersetzen", die im Web zitiert und mit anderen Resourcen verknüpft werden kann.
    Date
    15. 3.2011 19:21:22
    Source
    http://metadaten-twr.org/2011/01/19/skos-simple-knowledge-organisation-system/
    Theme
    Semantic Web
  16. Multilingual information management : current levels and future abilities. A report Commissioned by the US National Science Foundation and also delivered to the European Commission's Language Engineering Office and the US Defense Advanced Research Projects Agency, April 1999 (1999) 0.03
    0.0329042 = product of:
      0.09871259 = sum of:
        0.016146496 = weight(_text_:web in 6068) [ClassicSimilarity], result of:
          0.016146496 = score(doc=6068,freq=2.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.14422815 = fieldWeight in 6068, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=6068)
        0.014015876 = weight(_text_:information in 6068) [ClassicSimilarity], result of:
          0.014015876 = score(doc=6068,freq=18.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.23274568 = fieldWeight in 6068, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=6068)
        0.053511675 = weight(_text_:extraction in 6068) [ClassicSimilarity], result of:
          0.053511675 = score(doc=6068,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.26256397 = fieldWeight in 6068, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03125 = fieldNorm(doc=6068)
        0.015038553 = weight(_text_:system in 6068) [ClassicSimilarity], result of:
          0.015038553 = score(doc=6068,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.13919188 = fieldWeight in 6068, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03125 = fieldNorm(doc=6068)
      0.33333334 = coord(4/12)
    
    Abstract
    Over the past 50 years, a variety of language-related capabilities has been developed in machine translation, information retrieval, speech recognition, text summarization, and so on. These applications rest upon a set of core techniques such as language modeling, information extraction, parsing, generation, and multimedia planning and integration; and they involve methods using statistics, rules, grammars, lexicons, ontologies, training techniques, and so on. It is a puzzling fact that although all of this work deals with language in some form or other, the major applications have each developed a separate research field. For example, there is no reason why speech recognition techniques involving n-grams and hidden Markov models could not have been used in machine translation 15 years earlier than they were, or why some of the lexical and semantic insights from the subarea called Computational Linguistics are still not used in information retrieval.
    This picture will rapidly change. The twin challenges of massive information overload via the web and ubiquitous computers present us with an unavoidable task: developing techniques to handle multilingual and multi-modal information robustly and efficiently, with as high quality performance as possible. The most effective way for us to address such a mammoth task, and to ensure that our various techniques and applications fit together, is to start talking across the artificial research boundaries. Extending the current technologies will require integrating the various capabilities into multi-functional and multi-lingual natural language systems. However, at this time there is no clear vision of how these technologies could or should be assembled into a coherent framework. What would be involved in connecting a speech recognition system to an information retrieval engine, and then using machine translation and summarization software to process the retrieved text? How can traditional parsing and generation be enhanced with statistical techniques? What would be the effect of carefully crafted lexicons on traditional information retrieval? At which points should machine translation be interleaved within information retrieval systems to enable multilingual processing?
  17. Manguinhas, H.; Freire, N.; Machado, J.; Borbinha, J.: Supporting multilingual bibliographic resource discovery with Functional Requirements for Bibliographic Records (2012) 0.03
    0.03270465 = product of:
      0.1308186 = sum of:
        0.04513083 = weight(_text_:web in 133) [ClassicSimilarity], result of:
          0.04513083 = score(doc=133,freq=10.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.40312994 = fieldWeight in 133, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=133)
        0.06688959 = weight(_text_:extraction in 133) [ClassicSimilarity], result of:
          0.06688959 = score(doc=133,freq=2.0), product of:
            0.20380433 = queryWeight, product of:
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.03430388 = queryNorm
            0.32820496 = fieldWeight in 133, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.941145 = idf(docFreq=315, maxDocs=44218)
              0.0390625 = fieldNorm(doc=133)
        0.018798191 = weight(_text_:system in 133) [ClassicSimilarity], result of:
          0.018798191 = score(doc=133,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.17398985 = fieldWeight in 133, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=133)
      0.25 = coord(3/12)
    
    Abstract
    This paper describes an experiment exploring the hypothesis that innovative application of the Functional Require-ments for Bibliographic Records (FRBR) principles can complement traditional bibliographic resource discovery systems in order to improve the user experience. A specialized service was implemented that, when given a plain list of results from a regular online catalogue, was able to process, enrich and present that list in a more relevant way for the user. This service pre-processes the records of a traditional online catalogue in order to build a semantic structure following the FRBR model. The service also explores web search features that have been revolutionizing the way users conceptualize resource discovery, such as relevance ranking and metasearching. This work was developed in the context of the TELPlus project. We processed nearly one hundred thousand bibliographic and authority records, in multiple languages, and originating from twelve European na-tional libraries. This paper describes the architecture of the service and the main challenges faced, especially concerning the extraction and linking of the relevant FRBR entities from the bibliographic metadata produced by the libraries. The service was evaluated by end users, who filled out a questionnaire after using a traditional online catalogue and the new service, both with the same bibliographic collection. The analysis of the results supports the hypothesis that FRBR can be implemented for re-source discovery in a non-intrusive way, reusing the data of any existing traditional bibliographic system.
    Content
    Beitrag eines Schwerpunktthemas: Semantic Web and Reasoning for Cultural Heritage and Digital Libraries: Vgl.: http://www.semantic-web-journal.net/content/supporting-multilingual-bibliographic-resource-discovery-functional-requirements-bibliograph http://www.semantic-web-journal.net/sites/default/files/swj145_2.pdf.
    Source
    Semantic Web journal. 3(2012) no.1, S.3-21
  18. Tudhope, D.: Knowledge Organization System Services : brief review of NKOS activities and possibility of KOS registries (2007) 0.03
    0.030360341 = product of:
      0.121441364 = sum of:
        0.048439488 = weight(_text_:web in 100) [ClassicSimilarity], result of:
          0.048439488 = score(doc=100,freq=2.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.43268442 = fieldWeight in 100, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.09375 = fieldNorm(doc=100)
        0.04511566 = weight(_text_:system in 100) [ClassicSimilarity], result of:
          0.04511566 = score(doc=100,freq=2.0), product of:
            0.10804188 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03430388 = queryNorm
            0.41757566 = fieldWeight in 100, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.09375 = fieldNorm(doc=100)
        0.027886212 = product of:
          0.055772424 = sum of:
            0.055772424 = weight(_text_:22 in 100) [ClassicSimilarity], result of:
              0.055772424 = score(doc=100,freq=2.0), product of:
                0.120126344 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03430388 = queryNorm
                0.46428138 = fieldWeight in 100, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=100)
          0.5 = coord(1/2)
      0.25 = coord(3/12)
    
    Date
    22. 9.2007 15:41:14
    Theme
    Semantic Web
  19. Koch, T.: Searching the Web : systematic overview over indexes (1995) 0.03
    0.030338569 = product of:
      0.18203141 = sum of:
        0.06850378 = weight(_text_:web in 3169) [ClassicSimilarity], result of:
          0.06850378 = score(doc=3169,freq=4.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.6119082 = fieldWeight in 3169, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.09375 = fieldNorm(doc=3169)
        0.113527626 = weight(_text_:suche in 3169) [ClassicSimilarity], result of:
          0.113527626 = score(doc=3169,freq=2.0), product of:
            0.17138755 = queryWeight, product of:
              4.996156 = idf(docFreq=812, maxDocs=44218)
              0.03430388 = queryNorm
            0.662403 = fieldWeight in 3169, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.996156 = idf(docFreq=812, maxDocs=44218)
              0.09375 = fieldNorm(doc=3169)
      0.16666667 = coord(2/12)
    
    Abstract
    Vorstellung von Hilfsmitteln zur inhaltlichen Suche von Angeboten im Internet
    Object
    Nordic Web Index
  20. Hüsken, P.: Information Retrieval im Semantic Web (2006) 0.03
    0.030207884 = product of:
      0.120831534 = sum of:
        0.054157 = weight(_text_:web in 4333) [ClassicSimilarity], result of:
          0.054157 = score(doc=4333,freq=10.0), product of:
            0.111951075 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03430388 = queryNorm
            0.48375595 = fieldWeight in 4333, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4333)
        0.009910721 = weight(_text_:information in 4333) [ClassicSimilarity], result of:
          0.009910721 = score(doc=4333,freq=4.0), product of:
            0.060219705 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03430388 = queryNorm
            0.16457605 = fieldWeight in 4333, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=4333)
        0.056763813 = weight(_text_:suche in 4333) [ClassicSimilarity], result of:
          0.056763813 = score(doc=4333,freq=2.0), product of:
            0.17138755 = queryWeight, product of:
              4.996156 = idf(docFreq=812, maxDocs=44218)
              0.03430388 = queryNorm
            0.3312015 = fieldWeight in 4333, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.996156 = idf(docFreq=812, maxDocs=44218)
              0.046875 = fieldNorm(doc=4333)
      0.25 = coord(3/12)
    
    Abstract
    Das Semantic Web bezeichnet ein erweitertes World Wide Web (WWW), das die Bedeutung von präsentierten Inhalten in neuen standardisierten Sprachen wie RDF Schema und OWL modelliert. Diese Arbeit befasst sich mit dem Aspekt des Information Retrieval, d.h. es wird untersucht, in wie weit Methoden der Informationssuche sich auf modelliertes Wissen übertragen lassen. Die kennzeichnenden Merkmale von IR-Systemen wie vage Anfragen sowie die Unterstützung unsicheren Wissens werden im Kontext des Semantic Web behandelt. Im Fokus steht die Suche nach Fakten innerhalb einer Wissensdomäne, die entweder explizit modelliert sind oder implizit durch die Anwendung von Inferenz abgeleitet werden können. Aufbauend auf der an der Universität Duisburg-Essen entwickelten Retrievalmaschine PIRE wird die Anwendung unsicherer Inferenz mit probabilistischer Prädikatenlogik (pDatalog) implementiert.
    Theme
    Semantic Web

Years

Languages

Types

  • a 500
  • r 25
  • i 24
  • m 21
  • x 21
  • s 19
  • n 16
  • p 6
  • b 5
  • More… Less…

Themes