Search (87 results, page 1 of 5)

  • × type_ss:"x"
  1. Verwer, K.: Freiheit und Verantwortung bei Hans Jonas (2011) 0.39
    0.39391923 = product of:
      0.91914487 = sum of:
        0.13130641 = product of:
          0.3939192 = sum of:
            0.3939192 = weight(_text_:3a in 973) [ClassicSimilarity], result of:
              0.3939192 = score(doc=973,freq=2.0), product of:
                0.35045066 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.041336425 = queryNorm
                1.1240361 = fieldWeight in 973, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.09375 = fieldNorm(doc=973)
          0.33333334 = coord(1/3)
        0.3939192 = weight(_text_:2f in 973) [ClassicSimilarity], result of:
          0.3939192 = score(doc=973,freq=2.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            1.1240361 = fieldWeight in 973, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.09375 = fieldNorm(doc=973)
        0.3939192 = weight(_text_:2f in 973) [ClassicSimilarity], result of:
          0.3939192 = score(doc=973,freq=2.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            1.1240361 = fieldWeight in 973, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.09375 = fieldNorm(doc=973)
      0.42857143 = coord(3/7)
    
    Content
    Vgl.: http%3A%2F%2Fcreativechoice.org%2Fdoc%2FHansJonas.pdf&usg=AOvVaw1TM3teaYKgABL5H9yoIifA&opi=89978449.
  2. Gabler, S.: Vergabe von DDC-Sachgruppen mittels eines Schlagwort-Thesaurus (2021) 0.23
    0.23353532 = product of:
      0.4086868 = sum of:
        0.054711 = product of:
          0.164133 = sum of:
            0.164133 = weight(_text_:3a in 1000) [ClassicSimilarity], result of:
              0.164133 = score(doc=1000,freq=2.0), product of:
                0.35045066 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.041336425 = queryNorm
                0.46834838 = fieldWeight in 1000, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1000)
          0.33333334 = coord(1/3)
        0.164133 = weight(_text_:2f in 1000) [ClassicSimilarity], result of:
          0.164133 = score(doc=1000,freq=2.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            0.46834838 = fieldWeight in 1000, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1000)
        0.164133 = weight(_text_:2f in 1000) [ClassicSimilarity], result of:
          0.164133 = score(doc=1000,freq=2.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            0.46834838 = fieldWeight in 1000, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1000)
        0.025709787 = product of:
          0.051419575 = sum of:
            0.051419575 = weight(_text_:studies in 1000) [ClassicSimilarity], result of:
              0.051419575 = score(doc=1000,freq=4.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.3117402 = fieldWeight in 1000, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1000)
          0.5 = coord(1/2)
      0.5714286 = coord(4/7)
    
    Content
    Master thesis Master of Science (Library and Information Studies) (MSc), Universität Wien. Advisor: Christoph Steiner. Vgl.: https://www.researchgate.net/publication/371680244_Vergabe_von_DDC-Sachgruppen_mittels_eines_Schlagwort-Thesaurus. DOI: 10.25365/thesis.70030. Vgl. dazu die Präsentation unter: https://www.google.com/url?sa=i&rct=j&q=&esrc=s&source=web&cd=&ved=0CAIQw7AJahcKEwjwoZzzytz_AhUAAAAAHQAAAAAQAg&url=https%3A%2F%2Fwiki.dnb.de%2Fdownload%2Fattachments%2F252121510%2FDA3%2520Workshop-Gabler.pdf%3Fversion%3D1%26modificationDate%3D1671093170000%26api%3Dv2&psig=AOvVaw0szwENK1or3HevgvIDOfjx&ust=1687719410889597&opi=89978449.
    Imprint
    Wien / Library and Information Studies : Universität
  3. Stojanovic, N.: Ontology-based Information Retrieval : methods and tools for cooperative query answering (2005) 0.18
    0.18338586 = product of:
      0.32092524 = sum of:
        0.0437688 = product of:
          0.1313064 = sum of:
            0.1313064 = weight(_text_:3a in 701) [ClassicSimilarity], result of:
              0.1313064 = score(doc=701,freq=2.0), product of:
                0.35045066 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.041336425 = queryNorm
                0.3746787 = fieldWeight in 701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
          0.33333334 = coord(1/3)
        0.1313064 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.1313064 = score(doc=701,freq=2.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.1313064 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.1313064 = score(doc=701,freq=2.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.014543652 = product of:
          0.029087303 = sum of:
            0.029087303 = weight(_text_:studies in 701) [ClassicSimilarity], result of:
              0.029087303 = score(doc=701,freq=2.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.17634688 = fieldWeight in 701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
          0.5 = coord(1/2)
      0.5714286 = coord(4/7)
    
    Abstract
    By the explosion of possibilities for a ubiquitous content production, the information overload problem reaches the level of complexity which cannot be managed by traditional modelling approaches anymore. Due to their pure syntactical nature traditional information retrieval approaches did not succeed in treating content itself (i.e. its meaning, and not its representation). This leads to a very low usefulness of the results of a retrieval process for a user's task at hand. In the last ten years ontologies have been emerged from an interesting conceptualisation paradigm to a very promising (semantic) modelling technology, especially in the context of the Semantic Web. From the information retrieval point of view, ontologies enable a machine-understandable form of content description, such that the retrieval process can be driven by the meaning of the content. However, the very ambiguous nature of the retrieval process in which a user, due to the unfamiliarity with the underlying repository and/or query syntax, just approximates his information need in a query, implies a necessity to include the user in the retrieval process more actively in order to close the gap between the meaning of the content and the meaning of a user's query (i.e. his information need). This thesis lays foundation for such an ontology-based interactive retrieval process, in which the retrieval system interacts with a user in order to conceptually interpret the meaning of his query, whereas the underlying domain ontology drives the conceptualisation process. In that way the retrieval process evolves from a query evaluation process into a highly interactive cooperation between a user and the retrieval system, in which the system tries to anticipate the user's information need and to deliver the relevant content proactively. Moreover, the notion of content relevance for a user's query evolves from a content dependent artefact to the multidimensional context-dependent structure, strongly influenced by the user's preferences. This cooperation process is realized as the so-called Librarian Agent Query Refinement Process. In order to clarify the impact of an ontology on the retrieval process (regarding its complexity and quality), a set of methods and tools for different levels of content and query formalisation is developed, ranging from pure ontology-based inferencing to keyword-based querying in which semantics automatically emerges from the results. Our evaluation studies have shown that the possibilities to conceptualize a user's information need in the right manner and to interpret the retrieval results accordingly are key issues for realizing much more meaningful information retrieval systems.
    Content
    Vgl.: http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F1627&ei=tAtYUYrBNoHKtQb3l4GYBw&usg=AFQjCNHeaxKkKU3-u54LWxMNYGXaaDLCGw&sig2=8WykXWQoDKjDSdGtAakH2Q&bvm=bv.44442042,d.Yms.
  4. Xiong, C.: Knowledge based text representations for information retrieval (2016) 0.18
    0.17792545 = product of:
      0.41515937 = sum of:
        0.0437688 = product of:
          0.1313064 = sum of:
            0.1313064 = weight(_text_:3a in 5820) [ClassicSimilarity], result of:
              0.1313064 = score(doc=5820,freq=2.0), product of:
                0.35045066 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.041336425 = queryNorm
                0.3746787 = fieldWeight in 5820, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5820)
          0.33333334 = coord(1/3)
        0.18569529 = weight(_text_:2f in 5820) [ClassicSimilarity], result of:
          0.18569529 = score(doc=5820,freq=4.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            0.5298757 = fieldWeight in 5820, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
        0.18569529 = weight(_text_:2f in 5820) [ClassicSimilarity], result of:
          0.18569529 = score(doc=5820,freq=4.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            0.5298757 = fieldWeight in 5820, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
      0.42857143 = coord(3/7)
    
    Content
    Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Language and Information Technologies. Vgl.: https%3A%2F%2Fwww.cs.cmu.edu%2F~cx%2Fpapers%2Fknowledge_based_text_representation.pdf&usg=AOvVaw0SaTSvhWLTh__Uz_HtOtl3.
  5. Huo, W.: Automatic multi-word term extraction and its application to Web-page summarization (2012) 0.18
    0.17602317 = product of:
      0.41072074 = sum of:
        0.1969596 = weight(_text_:2f in 563) [ClassicSimilarity], result of:
          0.1969596 = score(doc=563,freq=2.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            0.56201804 = fieldWeight in 563, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=563)
        0.1969596 = weight(_text_:2f in 563) [ClassicSimilarity], result of:
          0.1969596 = score(doc=563,freq=2.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            0.56201804 = fieldWeight in 563, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=563)
        0.016801544 = product of:
          0.033603087 = sum of:
            0.033603087 = weight(_text_:22 in 563) [ClassicSimilarity], result of:
              0.033603087 = score(doc=563,freq=2.0), product of:
                0.14475311 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.041336425 = queryNorm
                0.23214069 = fieldWeight in 563, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=563)
          0.5 = coord(1/2)
      0.42857143 = coord(3/7)
    
    Content
    A Thesis presented to The University of Guelph In partial fulfilment of requirements for the degree of Master of Science in Computer Science. Vgl. Unter: http://www.inf.ufrgs.br%2F~ceramisch%2Fdownload_files%2Fpublications%2F2009%2Fp01.pdf.
    Date
    10. 1.2013 19:22:47
  6. Farazi, M.: Faceted lightweight ontologies : a formalization and some experiments (2010) 0.16
    0.16413301 = product of:
      0.382977 = sum of:
        0.054711 = product of:
          0.164133 = sum of:
            0.164133 = weight(_text_:3a in 4997) [ClassicSimilarity], result of:
              0.164133 = score(doc=4997,freq=2.0), product of:
                0.35045066 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.041336425 = queryNorm
                0.46834838 = fieldWeight in 4997, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4997)
          0.33333334 = coord(1/3)
        0.164133 = weight(_text_:2f in 4997) [ClassicSimilarity], result of:
          0.164133 = score(doc=4997,freq=2.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            0.46834838 = fieldWeight in 4997, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4997)
        0.164133 = weight(_text_:2f in 4997) [ClassicSimilarity], result of:
          0.164133 = score(doc=4997,freq=2.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            0.46834838 = fieldWeight in 4997, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4997)
      0.42857143 = coord(3/7)
    
    Content
    PhD Dissertation at International Doctorate School in Information and Communication Technology. Vgl.: https%3A%2F%2Fcore.ac.uk%2Fdownload%2Fpdf%2F150083013.pdf&usg=AOvVaw2n-qisNagpyT0lli_6QbAQ.
  7. Shala, E.: ¬Die Autonomie des Menschen und der Maschine : gegenwärtige Definitionen von Autonomie zwischen philosophischem Hintergrund und technologischer Umsetzbarkeit (2014) 0.16
    0.16413301 = product of:
      0.382977 = sum of:
        0.054711 = product of:
          0.164133 = sum of:
            0.164133 = weight(_text_:3a in 4388) [ClassicSimilarity], result of:
              0.164133 = score(doc=4388,freq=2.0), product of:
                0.35045066 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.041336425 = queryNorm
                0.46834838 = fieldWeight in 4388, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4388)
          0.33333334 = coord(1/3)
        0.164133 = weight(_text_:2f in 4388) [ClassicSimilarity], result of:
          0.164133 = score(doc=4388,freq=2.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            0.46834838 = fieldWeight in 4388, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4388)
        0.164133 = weight(_text_:2f in 4388) [ClassicSimilarity], result of:
          0.164133 = score(doc=4388,freq=2.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            0.46834838 = fieldWeight in 4388, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4388)
      0.42857143 = coord(3/7)
    
    Footnote
    Vgl. unter: https://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=2&cad=rja&uact=8&ved=2ahUKEwizweHljdbcAhVS16QKHXcFD9QQFjABegQICRAB&url=https%3A%2F%2Fwww.researchgate.net%2Fpublication%2F271200105_Die_Autonomie_des_Menschen_und_der_Maschine_-_gegenwartige_Definitionen_von_Autonomie_zwischen_philosophischem_Hintergrund_und_technologischer_Umsetzbarkeit_Redigierte_Version_der_Magisterarbeit_Karls&usg=AOvVaw06orrdJmFF2xbCCp_hL26q.
  8. Piros, A.: Az ETO-jelzetek automatikus interpretálásának és elemzésének kérdései (2018) 0.16
    0.16413301 = product of:
      0.382977 = sum of:
        0.054711 = product of:
          0.164133 = sum of:
            0.164133 = weight(_text_:3a in 855) [ClassicSimilarity], result of:
              0.164133 = score(doc=855,freq=2.0), product of:
                0.35045066 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.041336425 = queryNorm
                0.46834838 = fieldWeight in 855, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=855)
          0.33333334 = coord(1/3)
        0.164133 = weight(_text_:2f in 855) [ClassicSimilarity], result of:
          0.164133 = score(doc=855,freq=2.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            0.46834838 = fieldWeight in 855, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=855)
        0.164133 = weight(_text_:2f in 855) [ClassicSimilarity], result of:
          0.164133 = score(doc=855,freq=2.0), product of:
            0.35045066 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041336425 = queryNorm
            0.46834838 = fieldWeight in 855, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=855)
      0.42857143 = coord(3/7)
    
    Content
    Vgl. auch: New automatic interpreter for complex UDC numbers. Unter: <https%3A%2F%2Fudcc.org%2Ffiles%2FAttilaPiros_EC_36-37_2014-2015.pdf&usg=AOvVaw3kc9CwDDCWP7aArpfjrs5b>
  9. Geisriegler, E.: Enriching electronic texts with semantic metadata : a use case for the historical Newspaper Collection ANNO (Austrian Newspapers Online) of the Austrian National Libraryhek (2012) 0.04
    0.036223195 = product of:
      0.12678118 = sum of:
        0.062419478 = weight(_text_:case in 595) [ClassicSimilarity], result of:
          0.062419478 = score(doc=595,freq=4.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.34346986 = fieldWeight in 595, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.0390625 = fieldNorm(doc=595)
        0.06436171 = sum of:
          0.03635913 = weight(_text_:studies in 595) [ClassicSimilarity], result of:
            0.03635913 = score(doc=595,freq=2.0), product of:
              0.16494368 = queryWeight, product of:
                3.9902744 = idf(docFreq=2222, maxDocs=44218)
                0.041336425 = queryNorm
              0.22043361 = fieldWeight in 595, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.9902744 = idf(docFreq=2222, maxDocs=44218)
                0.0390625 = fieldNorm(doc=595)
          0.028002575 = weight(_text_:22 in 595) [ClassicSimilarity], result of:
            0.028002575 = score(doc=595,freq=2.0), product of:
              0.14475311 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041336425 = queryNorm
              0.19345059 = fieldWeight in 595, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=595)
      0.2857143 = coord(2/7)
    
    Abstract
    Die vorliegende Master Thesis setzt sich mit der Frage nach Möglichkeiten der Anreicherung historischer Zeitungen mit semantischen Metadaten auseinander. Sie möchte außerdem analysieren, welcher Nutzen für vor allem geisteswissenschaftlich Forschende, durch die Anreicherung mit zusätzlichen Informationsquellen entsteht. Nach der Darstellung der Entwicklung der interdisziplinären 'Digital Humanities', wurde für die digitale Sammlung historischer Zeitungen (ANNO AustriaN Newspapers Online) der Österreichischen Nationalbibliothek ein Use Case entwickelt, bei dem 'Named Entities' (Personen, Orte, Organisationen und Daten) in ausgewählten Zeitungsausgaben manuell annotiert wurden. Methodisch wurde das Kodieren mit 'TEI', einem Dokumentenformat zur Kodierung und zum Austausch von Texten durchgeführt. Zusätzlich wurden zu allen annotierten 'Named Entities' Einträge in externen Datenbanken wie Wikipedia, Wikipedia Personensuche, der ehemaligen Personennamen- und Schlagwortnormdatei (jetzt Gemeinsame Normdatei GND), VIAF und dem Bildarchiv Austria gesucht und gegebenenfalls verlinkt. Eine Beschreibung der Ergebnisse des manuellen Annotierens der Zeitungsseiten schließt diesen Teil der Arbeit ab. In einem weiteren Abschnitt werden die Ergebnisse des manuellen Annotierens mit jenen Ergebnissen, die automatisch mit dem German NER (Named Entity Recognition) generiert wurden, verglichen und in ihrer Genauigkeit analysiert. Abschließend präsentiert die Arbeit einige Best Practice-Beispiele kodierter und angereicherter Zeitungsseiten, um den zusätzlichen Nutzen durch die Auszeichnung der 'Named Entities' und durch die Verlinkung mit externen Informationsquellen für die BenützerInnen darzustellen.
    Date
    3. 2.2013 18:00:22
    Footnote
    Wien, Univ., Lehrgang Library and Information Studies, Master-Thesis, 2012.
  10. Castellanos Ardila, J.P.: Investigation of an OSLC-domain targeting ISO 26262 : focus on the left side of the software V-model (2016) 0.03
    0.030944746 = product of:
      0.10830661 = sum of:
        0.029351516 = weight(_text_:management in 5819) [ClassicSimilarity], result of:
          0.029351516 = score(doc=5819,freq=4.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.21066327 = fieldWeight in 5819, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.03125 = fieldNorm(doc=5819)
        0.07895509 = weight(_text_:case in 5819) [ClassicSimilarity], result of:
          0.07895509 = score(doc=5819,freq=10.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.43445885 = fieldWeight in 5819, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.03125 = fieldNorm(doc=5819)
      0.2857143 = coord(2/7)
    
    Abstract
    Industries have adopted a standardized set of practices for developing their products. In the automotive domain, the provision of safety-compliant systems is guided by ISO 26262, a standard that specifies a set of requirements and recommendations for developing automotive safety-critical systems. For being in compliance with ISO 26262, the safety lifecycle proposed by the standard must be included in the development process of a vehicle. Besides, a safety case that shows that the system is acceptably safe has to be provided. The provision of a safety case implies the execution of a precise documentation process. This process makes sure that the work products are available and traceable. Further, the documentation management is defined in the standard as a mandatory activity and guidelines are proposed/imposed for its elaboration. It would be appropriate to point out that a well-documented safety lifecycle will provide the necessary inputs for the generation of an ISO 26262-compliant safety case. The OSLC (Open Services for Lifecycle Collaboration) standard and the maturing stack of semantic web technologies represent a promising integration platform for enabling semantic interoperability between the tools involved in the safety lifecycle. Tools for requirements, architecture, development management, among others, are expected to interact and shared data with the help of domains specifications created in OSLC. This thesis proposes the creation of an OSLC tool-chain infrastructure for sharing safety-related information, where fragments of safety information can be generated. The steps carried out during the elaboration of this master thesis consist in the identification, representation, and shaping of the RDF resources needed for the creation of a safety case. The focus of the thesis is limited to a tiny portion of the ISO 26262 left-hand side of the V-model, more exactly part 6 clause 8 of the standard: Software unit design and implementation. Regardless of the use of a restricted portion of the standard during the execution of this thesis, the findings can be extended to other parts, and the conclusions can be generalize. This master thesis is considered one of the first steps towards the provision of an OSLC-based and ISO 26262-compliant methodological approach for representing and shaping the work products resulting from the execution of the safety lifecycle, documentation required in the conformation of an ISO-compliant safety case.
  11. Schwarz, K.: Domain model enhanced search : a comparison of taxonomy, thesaurus and ontology (2005) 0.02
    0.021629145 = product of:
      0.075702004 = sum of:
        0.061158355 = weight(_text_:case in 4569) [ClassicSimilarity], result of:
          0.061158355 = score(doc=4569,freq=6.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.3365304 = fieldWeight in 4569, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.03125 = fieldNorm(doc=4569)
        0.014543652 = product of:
          0.029087303 = sum of:
            0.029087303 = weight(_text_:studies in 4569) [ClassicSimilarity], result of:
              0.029087303 = score(doc=4569,freq=2.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.17634688 = fieldWeight in 4569, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4569)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    The results of this thesis are intended to support the information architect in designing a solution for improved search in a corporate environment. Specifically we have examined the type of search problems that require a domain model to enhance the search process. There are several approaches to modeling a domain. We have considered 3 different types of domain modeling schemes; taxonomy, thesaurus and ontology. The intention is to support the information architect in making an informed choice between one or more of these schemes. In our opinion the main criteria for this choice are the modeling characteristics of a scheme and the suitability for application in the search process. The second chapter is a discussion of modeling characteristics of each scheme, followed by a comparison between them. This should give an information architect an idea of which aspects of a domain can be modeled with each scheme. What is missing here is an indication of the effort required to model a domain with each scheme. There are too many factors that influence the amount of required effort, ranging from measurable factors like domain size and resource characteristics to cultural matters such as the willingness to share knowledge and the existence of a project champion in the team to keep the project running. The third chapter shows what role domain models can play in each part of the search process. This gives an idea of the problems that domain models can solve. We have split the search process into individual parts to show that domain models can be applied very differently in the process. The fourth chapter makes recommendations about the suitability of each individualdomain modeling scheme for improving search. Each scheme has particular characteristics that make it especially suitable for a domain or a search problem. In the appendix each case study is described in detail. These descriptions are intended to serve as a benchmark. The current problem of the enterprise can be compared to those described to see which case study is most similar, which solution was chosen, which problems arose and how they were dealt with. An important issue that we have not touched upon in this thesis is that of maintenance. The real problems of a domain model are revealed when it is applied in a search system and its deficits and wrong assumptions become clear. Adaptation and maintenance are always required. Unfortunately we have not been able to glean sufficient information about maintenance issues from our case studies to draw any meaningful conclusions.
  12. Slavic-Overfield, A.: Classification management and use in a networked environment : the case of the Universal Decimal Classification (2005) 0.02
    0.01847466 = product of:
      0.06466131 = sum of:
        0.029351516 = weight(_text_:management in 2191) [ClassicSimilarity], result of:
          0.029351516 = score(doc=2191,freq=4.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.21066327 = fieldWeight in 2191, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.03125 = fieldNorm(doc=2191)
        0.03530979 = weight(_text_:case in 2191) [ClassicSimilarity], result of:
          0.03530979 = score(doc=2191,freq=2.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.1942959 = fieldWeight in 2191, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.03125 = fieldNorm(doc=2191)
      0.2857143 = coord(2/7)
    
    Abstract
    In the Internet information space, advanced information retrieval (IR) methods and automatic text processing are used in conjunction with traditional knowledge organization systems (KOS). New information technology provides a platform for better KOS publishing, exploitation and sharing both for human and machine use. Networked KOS services are now being planned and developed as powerful tools for resource discovery. They will enable automatic contextualisation, interpretation and query matching to different indexing languages. The Semantic Web promises to be an environment in which the quality of semantic relationships in bibliographic classification systems can be fully exploited. Their use in the networked environment is, however, limited by the fact that they are not prepared or made available for advanced machine processing. The UDC was chosen for this research because of its widespread use and its long-term presence in online information retrieval systems. It was also the first system to be used for the automatic classification of Internet resources, and the first to be made available as a classification tool on the Web. The objective of this research is to establish the advantages of using UDC for information retrieval in a networked environment, to highlight the problems of automation and classification exchange, and to offer possible solutions. The first research question was is there enough evidence of the use of classification on the Internet to justify further development with this particular environment in mind? The second question is what are the automation requirements for the full exploitation of UDC and its exchange? The third question is which areas are in need of improvement and what specific recommendations can be made for implementing the UDC in a networked environment? A summary of changes required in the management and development of the UDC to facilitate its full adaptation for future use is drawn from this analysis.
  13. Pepper, S.: ¬The typology and semantics of binominal lexemes : noun-noun compounds and their functional equivalents (2020) 0.01
    0.014243842 = product of:
      0.049853444 = sum of:
        0.03530979 = weight(_text_:case in 104) [ClassicSimilarity], result of:
          0.03530979 = score(doc=104,freq=2.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.1942959 = fieldWeight in 104, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.03125 = fieldNorm(doc=104)
        0.014543652 = product of:
          0.029087303 = sum of:
            0.029087303 = weight(_text_:studies in 104) [ClassicSimilarity], result of:
              0.029087303 = score(doc=104,freq=2.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.17634688 = fieldWeight in 104, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.03125 = fieldNorm(doc=104)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    The dissertation establishes 'binominal lexeme' as a comparative concept and discusses its cross-linguistic typology and semantics. Informally, a binominal lexeme is a noun-noun compound or functional equivalent; more precisely, it is a lexical item that consists primarily of two thing-morphs between which there exists an unstated semantic relation. Examples of binominals include Mandarin Chinese ?? (tielù) [iron road], French chemin de fer [way of iron] and Russian ???????? ?????? (zeleznaja doroga) [iron:adjz road]. All of these combine a word denoting 'iron' and a word denoting 'road' or 'way' to denote the meaning railway. In each case, the unstated semantic relation is one of composition: a railway is conceptualized as a road that is composed (or made) of iron. However, three different morphosyntactic strategies are employed: compounding, prepositional phrase and relational adjective. This study explores the range of such strategies used by a worldwide sample of 106 languages to express a set of 100 meanings from various semantic domains, resulting in a classification consisting of nine different morphosyntactic types. The semantic relations found in the data are also explored and a classification called the Hatcher-Bourque system is developed that operates at two levels of granularity, together with a tool for classifying binominals, the Bourquifier. The classification is extended to other subfields of language, including metonymy and lexical semantics, and beyond language to the domain of knowledge representation, resulting in a proposal for a general model of associative relations called the PHAB model. The many findings of the research include universals concerning the recruitment of anchoring nominal modification strategies, a method for comparing non-binary typologies, the non-universality (despite its predominance) of compounding, and a scale of frequencies for semantic relations which may provide insights into the associative nature of human thought.
    Imprint
    Oslo : University of Oslo / Faculty of Humanities / Department of Linguistics and Scandinavian Studies
  14. Jockel, S.: Entwicklung eines Kalkulationsmodells für Informationsdienstleistungen (1990) 0.01
    0.011859803 = product of:
      0.08301862 = sum of:
        0.08301862 = weight(_text_:management in 2751) [ClassicSimilarity], result of:
          0.08301862 = score(doc=2751,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.5958457 = fieldWeight in 2751, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.125 = fieldNorm(doc=2751)
      0.14285715 = coord(1/7)
    
    Theme
    Information Resources Management
  15. Ruhland, B.: Entwurf und Implementierung eines Relationalen Information-Retrieval-Systems über Datenbank-Management-Systeme (1991) 0.01
    0.011859803 = product of:
      0.08301862 = sum of:
        0.08301862 = weight(_text_:management in 2760) [ClassicSimilarity], result of:
          0.08301862 = score(doc=2760,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.5958457 = fieldWeight in 2760, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.125 = fieldNorm(doc=2760)
      0.14285715 = coord(1/7)
    
  16. Holschbach, K.: Vorgehensweise bei der Einführung eines Dokumenten-Management-Systems : eine projektbegleitende Untersuchung (1995) 0.01
    0.011859803 = product of:
      0.08301862 = sum of:
        0.08301862 = weight(_text_:management in 5954) [ClassicSimilarity], result of:
          0.08301862 = score(doc=5954,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.5958457 = fieldWeight in 5954, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.125 = fieldNorm(doc=5954)
      0.14285715 = coord(1/7)
    
  17. Wagner, M.: Vorgehensweise und Problematik der Informationsbeschaffung und -auswertung bei der Erstellung einer Konkurrenzanalyse (1995) 0.01
    0.011859803 = product of:
      0.08301862 = sum of:
        0.08301862 = weight(_text_:management in 5966) [ClassicSimilarity], result of:
          0.08301862 = score(doc=5966,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.5958457 = fieldWeight in 5966, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.125 = fieldNorm(doc=5966)
      0.14285715 = coord(1/7)
    
    Theme
    Information Resources Management
  18. Buß, M.: Unternehmenssprache in internationalen Unternehmen : Probleme des Informationstransfers in der internen Kommunikation (2005) 0.01
    0.011412744 = product of:
      0.039944604 = sum of:
        0.025943318 = weight(_text_:management in 1482) [ClassicSimilarity], result of:
          0.025943318 = score(doc=1482,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.18620178 = fieldWeight in 1482, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1482)
        0.0140012875 = product of:
          0.028002575 = sum of:
            0.028002575 = weight(_text_:22 in 1482) [ClassicSimilarity], result of:
              0.028002575 = score(doc=1482,freq=2.0), product of:
                0.14475311 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.041336425 = queryNorm
                0.19345059 = fieldWeight in 1482, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1482)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Date
    22. 5.2005 18:25:26
    Theme
    Information Resources Management
  19. Sauer, C.S.: Analyse von Webcommunities und Extraktion von Wissen aus Communitydaten für Case-Based Reasoning Systeme (2010) 0.01
    0.0109211365 = product of:
      0.07644795 = sum of:
        0.07644795 = weight(_text_:case in 2231) [ClassicSimilarity], result of:
          0.07644795 = score(doc=2231,freq=6.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.420663 = fieldWeight in 2231, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2231)
      0.14285715 = coord(1/7)
    
    Abstract
    Die vorliegende Arbeit befasst sich mit den Möglichkeiten der Informationsextraktion aus den Daten von Webcommunities und der Verwendung der extrahierten Informationen in Case-Based Reasoning- (CBR) Systemen. Im Rahmen der Arbeit wird auf die Entwicklung der Webcommunities im Zeitraum der letzten 30 Jahre eingegangen. Es wird eine Klassifikation der derzeitig anzutreffenden Webcommunities in verschiedene Kategorien von Webcommunities vorgenommen. Diese Klassifikation erfolgt hinsichtlich der Struktur, der technischen Mittel sowie der Interessen der Nutzer dieser Webcommunities. Aufbauend auf die vorgenommene Klassifikation von Webcommunities erfolgt eine Untersuchung der Eignung dieser Kategorien von Webcommunities zur Informationsextraktion im Kontext der Verwendung der extrahierten Informationen in CBR-Systemen. Im selben Kontext werden verschiedene Ansätze und Techniken der Informationsextraktion auf ihre Eignung zur Extraktion von Wissen speziell für die Wissenscontainer von CBR -Systeme geprüft. Aufbauend auf den dadurch gewonnenen Erkenntnissen wird, angelehnt an den Prozess der Knowledge Discovery in Databases, ein eigenes Prozessmodell der Wissensextraktion aus Webcommunities für CBR-Systeme entworfen. Im Zuge der näheren Betrachtung dieses Prozessmodells wird auf verschiedene, durch die beabsichtigte Verwendung der extrahierten Informationen in den vier Wissenscontainern des CBR bedingte, Anforderungen an NLP- sowie IE-Techniken, die zur Extraktion dieser Daten verwendet werden, eingegangen. Die in den theoretischen Betrachtungen erlangten Erkenntnisse werden dann dazu eingesetzt, eine Anwendung zur Informationsextraktion aus einer Webcommunity für ein CBR-System, in Form der Knowledge Extraction Workbench zu implementieren. Diese IEAnwendung arbeitet im Kontext des auf der SEASALT-Architektur aufbauenden Projektes docQuery. Die Realisierung dieser IE-Anwendung wird dokumentiert sowie die Extraktionsergebnisse der Anwendung hinsichtlich ihres Umfanges und ihrer Qualität evaluiert.
    Theme
    Case Based Reasoning
  20. Habermann, K.: Wissensrepräsentation im Rahmen von Wissensmanagement (1999) 0.01
    0.010377328 = product of:
      0.07264129 = sum of:
        0.07264129 = weight(_text_:management in 1515) [ClassicSimilarity], result of:
          0.07264129 = score(doc=1515,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.521365 = fieldWeight in 1515, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.109375 = fieldNorm(doc=1515)
      0.14285715 = coord(1/7)
    
    Theme
    Information Resources Management

Years

Languages

  • d 61
  • e 24
  • hu 1
  • pt 1
  • More… Less…

Types