Search (53 results, page 1 of 3)

  • × type_ss:"x"
  1. Xiong, C.: Knowledge based text representations for information retrieval (2016) 0.13
    0.12872586 = product of:
      0.17163447 = sum of:
        0.050406437 = product of:
          0.15121931 = sum of:
            0.15121931 = weight(_text_:3a in 5820) [ClassicSimilarity], result of:
              0.15121931 = score(doc=5820,freq=2.0), product of:
                0.4035973 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.047605187 = queryNorm
                0.3746787 = fieldWeight in 5820, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5820)
          0.33333334 = coord(1/3)
        0.09923182 = weight(_text_:space in 5820) [ClassicSimilarity], result of:
          0.09923182 = score(doc=5820,freq=6.0), product of:
            0.24842183 = queryWeight, product of:
              5.2183776 = idf(docFreq=650, maxDocs=44218)
              0.047605187 = queryNorm
            0.39944884 = fieldWeight in 5820, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              5.2183776 = idf(docFreq=650, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
        0.021996219 = product of:
          0.043992437 = sum of:
            0.043992437 = weight(_text_:model in 5820) [ClassicSimilarity], result of:
              0.043992437 = score(doc=5820,freq=4.0), product of:
                0.1830527 = queryWeight, product of:
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.047605187 = queryNorm
                0.24032663 = fieldWeight in 5820, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5820)
          0.5 = coord(1/2)
      0.75 = coord(3/4)
    
    Abstract
    The successes of information retrieval (IR) in recent decades were built upon bag-of-words representations. Effective as it is, bag-of-words is only a shallow text understanding; there is a limited amount of information for document ranking in the word space. This dissertation goes beyond words and builds knowledge based text representations, which embed the external and carefully curated information from knowledge bases, and provide richer and structured evidence for more advanced information retrieval systems. This thesis research first builds query representations with entities associated with the query. Entities' descriptions are used by query expansion techniques that enrich the query with explanation terms. Then we present a general framework that represents a query with entities that appear in the query, are retrieved by the query, or frequently show up in the top retrieved documents. A latent space model is developed to jointly learn the connections from query to entities and the ranking of documents, modeling the external evidence from knowledge bases and internal ranking features cooperatively. To further improve the quality of relevant entities, a defining factor of our query representations, we introduce learning to rank to entity search and retrieve better entities from knowledge bases. In the document representation part, this thesis research also moves one step forward with a bag-of-entities model, in which documents are represented by their automatic entity annotations, and the ranking is performed in the entity space.
    Content
    Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Language and Information Technologies. Vgl.: https%3A%2F%2Fwww.cs.cmu.edu%2F~cx%2Fpapers%2Fknowledge_based_text_representation.pdf&usg=AOvVaw0SaTSvhWLTh__Uz_HtOtl3.
  2. Ziemba, L.: Information retrieval with concept discovery in digital collections for agriculture and natural resources (2011) 0.12
    0.120062895 = product of:
      0.16008386 = sum of:
        0.08723867 = weight(_text_:vector in 4728) [ClassicSimilarity], result of:
          0.08723867 = score(doc=4728,freq=2.0), product of:
            0.30654848 = queryWeight, product of:
              6.439392 = idf(docFreq=191, maxDocs=44218)
              0.047605187 = queryNorm
            0.2845836 = fieldWeight in 4728, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.439392 = idf(docFreq=191, maxDocs=44218)
              0.03125 = fieldNorm(doc=4728)
        0.05729151 = weight(_text_:space in 4728) [ClassicSimilarity], result of:
          0.05729151 = score(doc=4728,freq=2.0), product of:
            0.24842183 = queryWeight, product of:
              5.2183776 = idf(docFreq=650, maxDocs=44218)
              0.047605187 = queryNorm
            0.23062189 = fieldWeight in 4728, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.2183776 = idf(docFreq=650, maxDocs=44218)
              0.03125 = fieldNorm(doc=4728)
        0.015553676 = product of:
          0.031107351 = sum of:
            0.031107351 = weight(_text_:model in 4728) [ClassicSimilarity], result of:
              0.031107351 = score(doc=4728,freq=2.0), product of:
                0.1830527 = queryWeight, product of:
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.047605187 = queryNorm
                0.16993658 = fieldWeight in 4728, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4728)
          0.5 = coord(1/2)
      0.75 = coord(3/4)
    
    Abstract
    The amount and complexity of information available in a digital form is already huge and new information is being produced every day. Retrieving information relevant to address a particular need becomes a significant issue. This work utilizes knowledge organization systems (KOS), such as thesauri and ontologies and applies information extraction (IE) and computational linguistics (CL) techniques to organize, manage and retrieve information stored in digital collections in the agricultural domain. Two real world applications of the approach have been developed and are available and actively used by the public. An ontology is used to manage the Water Conservation Digital Library holding a dynamic collection of various types of digital resources in the domain of urban water conservation in Florida, USA. The ontology based back-end powers a fully operational web interface, available at http://library.conservefloridawater.org. The system has demonstrated numerous benefits of the ontology application, including accurate retrieval of resources, information sharing and reuse, and has proved to effectively facilitate information management. The major difficulty encountered with the approach is that large and dynamic number of concepts makes it difficult to keep the ontology consistent and to accurately catalog resources manually. To address the aforementioned issues, a combination of IE and CL techniques, such as Vector Space Model and probabilistic parsing, with the use of Agricultural Thesaurus were adapted to automatically extract concepts important for each of the texts in the Best Management Practices (BMP) Publication Library--a collection of documents in the domain of agricultural BMPs in Florida available at http://lyra.ifas.ufl.edu/LIB. A new approach of domain-specific concept discovery with the use of Internet search engine was developed. Initial evaluation of the results indicates significant improvement in precision of information extraction. The approach presented in this work focuses on problems unique to agriculture and natural resources domain, such as domain specific concepts and vocabularies, but should be applicable to any collection of texts in digital format. It may be of potential interest for anyone who needs to effectively manage a collection of digital resources.
  3. Gordon, T.J.; Helmer-Hirschberg, O.: Report on a long-range forecasting study (1964) 0.08
    0.07553443 = product of:
      0.15106887 = sum of:
        0.11458302 = weight(_text_:space in 4204) [ClassicSimilarity], result of:
          0.11458302 = score(doc=4204,freq=2.0), product of:
            0.24842183 = queryWeight, product of:
              5.2183776 = idf(docFreq=650, maxDocs=44218)
              0.047605187 = queryNorm
            0.46124378 = fieldWeight in 4204, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.2183776 = idf(docFreq=650, maxDocs=44218)
              0.0625 = fieldNorm(doc=4204)
        0.036485836 = product of:
          0.07297167 = sum of:
            0.07297167 = weight(_text_:22 in 4204) [ClassicSimilarity], result of:
              0.07297167 = score(doc=4204,freq=4.0), product of:
                0.16670525 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047605187 = queryNorm
                0.4377287 = fieldWeight in 4204, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4204)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Description of an experimental trend-predicting exercise covering a time period as far as 50 years into the future. The Delphi technique is used in soliciting the opinions of experts in six areas: scientific breakthroughs, population growth, automation, space progress, probability and prevention of war, and future weapon systems. Possible objections to the approach are also discussed.
    Date
    22. 6.2018 13:24:08
    22. 6.2018 13:54:52
  4. Kiren, T.: ¬A clustering based indexing technique of modularized ontologies for information retrieval (2017) 0.05
    0.04696106 = product of:
      0.09392212 = sum of:
        0.081022434 = weight(_text_:space in 4399) [ClassicSimilarity], result of:
          0.081022434 = score(doc=4399,freq=4.0), product of:
            0.24842183 = queryWeight, product of:
              5.2183776 = idf(docFreq=650, maxDocs=44218)
              0.047605187 = queryNorm
            0.3261486 = fieldWeight in 4399, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.2183776 = idf(docFreq=650, maxDocs=44218)
              0.03125 = fieldNorm(doc=4399)
        0.012899691 = product of:
          0.025799382 = sum of:
            0.025799382 = weight(_text_:22 in 4399) [ClassicSimilarity], result of:
              0.025799382 = score(doc=4399,freq=2.0), product of:
                0.16670525 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047605187 = queryNorm
                0.15476047 = fieldWeight in 4399, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4399)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Modular ontologies are built in modular manner by combining modules from multiple relevant ontologies. Ontology heterogeneity also arises during modular ontology construction because multiple ontologies are being dealt with, during this process. Ontologies need to be aligned before using them for modular ontology construction. The existing approaches for ontology alignment compare all the concepts of each ontology to be aligned, hence not optimized in terms of time and search space utilization. A new indexing technique is proposed based on modular ontology. An efficient ontology alignment technique is proposed to solve the heterogeneity problem during the construction of modular ontology. Results are satisfactory as Precision and Recall are improved by (8%) and (10%) respectively. The value of Pearsons Correlation Coefficient for degree of similarity, time, search space requirement, precision and recall are close to 1 which shows that the results are significant. Further research can be carried out for using modular ontology based indexing technique for Multimedia Information Retrieval and Bio-Medical information retrieval.
    Date
    20. 1.2015 18:30:22
  5. Tzitzikas, Y.: Collaborative ontology-based information indexing and retrieval (2002) 0.04
    0.04419943 = product of:
      0.08839886 = sum of:
        0.05729151 = weight(_text_:space in 2281) [ClassicSimilarity], result of:
          0.05729151 = score(doc=2281,freq=2.0), product of:
            0.24842183 = queryWeight, product of:
              5.2183776 = idf(docFreq=650, maxDocs=44218)
              0.047605187 = queryNorm
            0.23062189 = fieldWeight in 2281, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.2183776 = idf(docFreq=650, maxDocs=44218)
              0.03125 = fieldNorm(doc=2281)
        0.031107351 = product of:
          0.062214702 = sum of:
            0.062214702 = weight(_text_:model in 2281) [ClassicSimilarity], result of:
              0.062214702 = score(doc=2281,freq=8.0), product of:
                0.1830527 = queryWeight, product of:
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.047605187 = queryNorm
                0.33987316 = fieldWeight in 2281, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2281)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    An information system like the Web is a continuously evolving system consisting of multiple heterogeneous information sources, covering a wide domain of discourse, and a huge number of users (human or software) with diverse characteristics and needs, that produce and consume information. The challenge nowadays is to build a scalable information infrastructure enabling the effective, accurate, content based retrieval of information, in a way that adapts to the characteristics and interests of the users. The aim of this work is to propose formally sound methods for building such an information network based on ontologies which are widely used and are easy to grasp by ordinary Web users. The main results of this work are: - A novel scheme for indexing and retrieving objects according to multiple aspects or facets. The proposed scheme is a faceted scheme enriched with a method for specifying the combinations of terms that are valid. We give a model-theoretic interpretation to this model and we provide mechanisms for inferring the valid combinations of terms. This inference service can be exploited for preventing errors during the indexing process, which is very important especially in the case where the indexing is done collaboratively by many users, and for deriving "complete" navigation trees suitable for browsing through the Web. The proposed scheme has several advantages over the hierarchical classification schemes currently employed by Web catalogs, namely, conceptual clarity (it is easier to understand), compactness (it takes less space), and scalability (the update operations can be formulated more easily and be performed more effciently). - A exible and effecient model for building mediators over ontology based information sources. The proposed mediators support several modes of query translation and evaluation which can accommodate various application needs and levels of answer quality. The proposed model can be used for providing users with customized views of Web catalogs. It can also complement the techniques for building mediators over relational sources so as to support approximate translation of partially ordered domain values.
  6. Verwer, K.: Freiheit und Verantwortung bei Hans Jonas (2011) 0.04
    0.037804827 = product of:
      0.15121931 = sum of:
        0.15121931 = product of:
          0.45365793 = sum of:
            0.45365793 = weight(_text_:3a in 973) [ClassicSimilarity], result of:
              0.45365793 = score(doc=973,freq=2.0), product of:
                0.4035973 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.047605187 = queryNorm
                1.1240361 = fieldWeight in 973, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.09375 = fieldNorm(doc=973)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Content
    Vgl.: http%3A%2F%2Fcreativechoice.org%2Fdoc%2FHansJonas.pdf&usg=AOvVaw1TM3teaYKgABL5H9yoIifA&opi=89978449.
  7. Huo, W.: Automatic multi-word term extraction and its application to Web-page summarization (2012) 0.02
    0.021340024 = product of:
      0.085360095 = sum of:
        0.085360095 = sum of:
          0.046661027 = weight(_text_:model in 563) [ClassicSimilarity], result of:
            0.046661027 = score(doc=563,freq=2.0), product of:
              0.1830527 = queryWeight, product of:
                3.845226 = idf(docFreq=2569, maxDocs=44218)
                0.047605187 = queryNorm
              0.25490487 = fieldWeight in 563, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.845226 = idf(docFreq=2569, maxDocs=44218)
                0.046875 = fieldNorm(doc=563)
          0.03869907 = weight(_text_:22 in 563) [ClassicSimilarity], result of:
            0.03869907 = score(doc=563,freq=2.0), product of:
              0.16670525 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.047605187 = queryNorm
              0.23214069 = fieldWeight in 563, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=563)
      0.25 = coord(1/4)
    
    Abstract
    In this thesis we propose three new word association measures for multi-word term extraction. We combine these association measures with LocalMaxs algorithm in our extraction model and compare the results of different multi-word term extraction methods. Our approach is language and domain independent and requires no training data. It can be applied to such tasks as text summarization, information retrieval, and document classification. We further explore the potential of using multi-word terms as an effective representation for general web-page summarization. We extract multi-word terms from human written summaries in a large collection of web-pages, and generate the summaries by aligning document words with these multi-word terms. Our system applies machine translation technology to learn the aligning process from a training set and focuses on selecting high quality multi-word terms from human written summaries to generate suitable results for web-page summarization.
    Date
    10. 1.2013 19:22:47
  8. Farazi, M.: Faceted lightweight ontologies : a formalization and some experiments (2010) 0.02
    0.015752012 = product of:
      0.06300805 = sum of:
        0.06300805 = product of:
          0.18902414 = sum of:
            0.18902414 = weight(_text_:3a in 4997) [ClassicSimilarity], result of:
              0.18902414 = score(doc=4997,freq=2.0), product of:
                0.4035973 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.047605187 = queryNorm
                0.46834838 = fieldWeight in 4997, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4997)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Content
    PhD Dissertation at International Doctorate School in Information and Communication Technology. Vgl.: https%3A%2F%2Fcore.ac.uk%2Fdownload%2Fpdf%2F150083013.pdf&usg=AOvVaw2n-qisNagpyT0lli_6QbAQ.
  9. Shala, E.: ¬Die Autonomie des Menschen und der Maschine : gegenwärtige Definitionen von Autonomie zwischen philosophischem Hintergrund und technologischer Umsetzbarkeit (2014) 0.02
    0.015752012 = product of:
      0.06300805 = sum of:
        0.06300805 = product of:
          0.18902414 = sum of:
            0.18902414 = weight(_text_:3a in 4388) [ClassicSimilarity], result of:
              0.18902414 = score(doc=4388,freq=2.0), product of:
                0.4035973 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.047605187 = queryNorm
                0.46834838 = fieldWeight in 4388, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4388)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Footnote
    Vgl. unter: https://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=2&cad=rja&uact=8&ved=2ahUKEwizweHljdbcAhVS16QKHXcFD9QQFjABegQICRAB&url=https%3A%2F%2Fwww.researchgate.net%2Fpublication%2F271200105_Die_Autonomie_des_Menschen_und_der_Maschine_-_gegenwartige_Definitionen_von_Autonomie_zwischen_philosophischem_Hintergrund_und_technologischer_Umsetzbarkeit_Redigierte_Version_der_Magisterarbeit_Karls&usg=AOvVaw06orrdJmFF2xbCCp_hL26q.
  10. Piros, A.: Az ETO-jelzetek automatikus interpretálásának és elemzésének kérdései (2018) 0.02
    0.015752012 = product of:
      0.06300805 = sum of:
        0.06300805 = product of:
          0.18902414 = sum of:
            0.18902414 = weight(_text_:3a in 855) [ClassicSimilarity], result of:
              0.18902414 = score(doc=855,freq=2.0), product of:
                0.4035973 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.047605187 = queryNorm
                0.46834838 = fieldWeight in 855, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=855)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Content
    Vgl. auch: New automatic interpreter for complex UDC numbers. Unter: <https%3A%2F%2Fudcc.org%2Ffiles%2FAttilaPiros_EC_36-37_2014-2015.pdf&usg=AOvVaw3kc9CwDDCWP7aArpfjrs5b>
  11. Gabler, S.: Vergabe von DDC-Sachgruppen mittels eines Schlagwort-Thesaurus (2021) 0.02
    0.015752012 = product of:
      0.06300805 = sum of:
        0.06300805 = product of:
          0.18902414 = sum of:
            0.18902414 = weight(_text_:3a in 1000) [ClassicSimilarity], result of:
              0.18902414 = score(doc=1000,freq=2.0), product of:
                0.4035973 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.047605187 = queryNorm
                0.46834838 = fieldWeight in 1000, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1000)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Content
    Master thesis Master of Science (Library and Information Studies) (MSc), Universität Wien. Advisor: Christoph Steiner. Vgl.: https://www.researchgate.net/publication/371680244_Vergabe_von_DDC-Sachgruppen_mittels_eines_Schlagwort-Thesaurus. DOI: 10.25365/thesis.70030. Vgl. dazu die Präsentation unter: https://www.google.com/url?sa=i&rct=j&q=&esrc=s&source=web&cd=&ved=0CAIQw7AJahcKEwjwoZzzytz_AhUAAAAAHQAAAAAQAg&url=https%3A%2F%2Fwiki.dnb.de%2Fdownload%2Fattachments%2F252121510%2FDA3%2520Workshop-Gabler.pdf%3Fversion%3D1%26modificationDate%3D1671093170000%26api%3Dv2&psig=AOvVaw0szwENK1or3HevgvIDOfjx&ust=1687719410889597&opi=89978449.
  12. Ammann, A.: Klassifikation dynamischer Wissensräume : multifaktorielle Wechselbeziehungen zur Generierung und Gestaltung konstellativer dynamischer und mehrdimensionaler Wissensräume mit einem Fokus der Anwendung in der Zahn-, Mund- und Kieferheilkunde am Beispiel der enossalen Implantologie (2012) 0.02
    0.015191706 = product of:
      0.060766824 = sum of:
        0.060766824 = weight(_text_:space in 1751) [ClassicSimilarity], result of:
          0.060766824 = score(doc=1751,freq=4.0), product of:
            0.24842183 = queryWeight, product of:
              5.2183776 = idf(docFreq=650, maxDocs=44218)
              0.047605187 = queryNorm
            0.24461144 = fieldWeight in 1751, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.2183776 = idf(docFreq=650, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1751)
      0.25 = coord(1/4)
    
    Abstract
    Eine Neuorientierung der Wissensarchitektur bei der Generierung und Gestaltung der konstellativen dynamischen und mehrdimensionalen Wissensräume wird eine zentrale Herausforderung in der digitalen Wissenswelt werden. Die <Einheit des Wissens> wird dabei nicht mehr als universeller interdisziplinärer Wissensraum mit dem Problem <lost within the space of knowledge> verstanden, sondern assistierende Netz- Avatare konfigurieren den erforderlichen Wissensraum durch die Orientierung des konstellativen Nutzerprofils selbstständig. Es wäre jedoch vermessen, hier den Anspruch eines neuen Ordnungsprinzips des Wissens zu erheben. Mit dieser Arbeit können aber Impulse für einen "semantischen Code des Wissens" gegeben werden, wie durch eine Klassifikation der Wissensräume ein neues Konzept zur "Einheit des Wissens" in einem Raum-Zeit-Gefüge entstehen kann. Dabei werden wir uns auch der fundamentalen Frage stellen müssen, wie sich eine neue Wissenskultur definiert, in der das Jahrtausende währende Kulturerbe der Schriftlichkeit des Wissens sich zu einer Flüchtigkeit des Wissens als Gebrauchsgut hin entwickelt. Es ist die gesellschaftliche Position der genuinen intellektuellen Leistung des Urhebers und Schöpfers einer Schriftlichkeit, die er an die "Wisdom of Community" abgibt. Parallel hierzu finden auch Divergenz-Prozesse statt, in der sich die Schriftlichkeit von dem Medium löst, um selbst laufend neue Konvergenz- Prozesse einzugehen, in denen nicht mehr der Inhalt das Medium bestimmt, sondern das Medium die charakteristischen Merkmale der Inhalte definiert.
    Im Teil A wird, neben dem menschlichen Urbedürfnis mit dem Streben nach dem Wissen, auf die vier epochalen Konvergenz-Zyklen mit ihren Kompetenzprofilen der Wissensordnungen im Wissenstransfer eingegangen. Insbesondere die Verschiebungen der Wissenschaftssprachen nehmen dabei einen erheblichen Einfluss auf die Abgrenzung der Klassifikationen zum impliziten, visuellen und explizitem Wissen. Daher werden den Äquivalenztypen im expliziten Wissensraum einer besonderen Aufmerksamkeit gewidmet, denn in unserer multilingualen Wissenslandschaft entstehen im Wissenstransfer zum Verfügungs-, Orientierungs- und Handlungswissen Artefakte, die auch auf die Gestaltung der Lernziel-Taxonomien einen Einfluss haben. Im Teil B werden zunächst die Arten, Merkmale und Klassifikationskonzepte des Wissens behandelt. Bei dem Versuch einer neuen Wissensordnung wird das kartesische / geodätische Koordinatensystem in ein Raum-Zeit-Gefüge gestellt, aus dem sich elf Wissensräume herauskristallisiert haben, die sowohl in ihren Definitionen, den damit verbundenen Ableitungen und Beispielen und einer Verortung im Wissensraum klassifiziert werden. Im Projekt <K-Space Visual Library in Dental Medicine> wird die problem- und aufgabenorientierte Navigation in den jeweiligen Wissensräumen erläutert und in der Diskussion die noch bevorstehenden Konvergenz-Forderungen der meist noch bestehenden proprietären digitalen Technologien und Programme skizziert, um diese bei der Modellierung der Wissensräume mit einzubeziehen.
  13. Slavic-Overfield, A.: Classification management and use in a networked environment : the case of the Universal Decimal Classification (2005) 0.01
    0.014322878 = product of:
      0.05729151 = sum of:
        0.05729151 = weight(_text_:space in 2191) [ClassicSimilarity], result of:
          0.05729151 = score(doc=2191,freq=2.0), product of:
            0.24842183 = queryWeight, product of:
              5.2183776 = idf(docFreq=650, maxDocs=44218)
              0.047605187 = queryNorm
            0.23062189 = fieldWeight in 2191, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.2183776 = idf(docFreq=650, maxDocs=44218)
              0.03125 = fieldNorm(doc=2191)
      0.25 = coord(1/4)
    
    Abstract
    In the Internet information space, advanced information retrieval (IR) methods and automatic text processing are used in conjunction with traditional knowledge organization systems (KOS). New information technology provides a platform for better KOS publishing, exploitation and sharing both for human and machine use. Networked KOS services are now being planned and developed as powerful tools for resource discovery. They will enable automatic contextualisation, interpretation and query matching to different indexing languages. The Semantic Web promises to be an environment in which the quality of semantic relationships in bibliographic classification systems can be fully exploited. Their use in the networked environment is, however, limited by the fact that they are not prepared or made available for advanced machine processing. The UDC was chosen for this research because of its widespread use and its long-term presence in online information retrieval systems. It was also the first system to be used for the automatic classification of Internet resources, and the first to be made available as a classification tool on the Web. The objective of this research is to establish the advantages of using UDC for information retrieval in a networked environment, to highlight the problems of automation and classification exchange, and to offer possible solutions. The first research question was is there enough evidence of the use of classification on the Internet to justify further development with this particular environment in mind? The second question is what are the automation requirements for the full exploitation of UDC and its exchange? The third question is which areas are in need of improvement and what specific recommendations can be made for implementing the UDC in a networked environment? A summary of changes required in the management and development of the UDC to facilitate its full adaptation for future use is drawn from this analysis.
  14. Stünkel, M.: Neuere Methoden der inhaltlichen Erschließung schöner Literatur in öffentlichen Bibliotheken (1986) 0.01
    0.012899691 = product of:
      0.051598765 = sum of:
        0.051598765 = product of:
          0.10319753 = sum of:
            0.10319753 = weight(_text_:22 in 5815) [ClassicSimilarity], result of:
              0.10319753 = score(doc=5815,freq=2.0), product of:
                0.16670525 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047605187 = queryNorm
                0.61904186 = fieldWeight in 5815, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=5815)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    4. 8.2006 21:35:22
  15. Stojanovic, N.: Ontology-based Information Retrieval : methods and tools for cooperative query answering (2005) 0.01
    0.012601609 = product of:
      0.050406437 = sum of:
        0.050406437 = product of:
          0.15121931 = sum of:
            0.15121931 = weight(_text_:3a in 701) [ClassicSimilarity], result of:
              0.15121931 = score(doc=701,freq=2.0), product of:
                0.4035973 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.047605187 = queryNorm
                0.3746787 = fieldWeight in 701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Content
    Vgl.: http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F1627&ei=tAtYUYrBNoHKtQb3l4GYBw&usg=AFQjCNHeaxKkKU3-u54LWxMNYGXaaDLCGw&sig2=8WykXWQoDKjDSdGtAakH2Q&bvm=bv.44442042,d.Yms.
  16. Menges, T.: Möglichkeiten und Grenzen der Übertragbarkeit eines Buches auf Hypertext am Beispiel einer französischen Grundgrammatik (Klein; Kleineidam) (1997) 0.01
    0.011287229 = product of:
      0.045148917 = sum of:
        0.045148917 = product of:
          0.09029783 = sum of:
            0.09029783 = weight(_text_:22 in 1496) [ClassicSimilarity], result of:
              0.09029783 = score(doc=1496,freq=2.0), product of:
                0.16670525 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047605187 = queryNorm
                0.5416616 = fieldWeight in 1496, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1496)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    22. 7.1998 18:23:25
  17. Schneider, A.: ¬Die Verzeichnung und sachliche Erschließung der Belletristik in Kaysers Bücherlexikon und im Schlagwortkatalog Georg/Ost (1980) 0.01
    0.011287229 = product of:
      0.045148917 = sum of:
        0.045148917 = product of:
          0.09029783 = sum of:
            0.09029783 = weight(_text_:22 in 5309) [ClassicSimilarity], result of:
              0.09029783 = score(doc=5309,freq=2.0), product of:
                0.16670525 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047605187 = queryNorm
                0.5416616 = fieldWeight in 5309, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=5309)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    5. 8.2006 13:07:22
  18. Sperling, R.: Anlage von Literaturreferenzen für Onlineressourcen auf einer virtuellen Lernplattform (2004) 0.01
    0.011287229 = product of:
      0.045148917 = sum of:
        0.045148917 = product of:
          0.09029783 = sum of:
            0.09029783 = weight(_text_:22 in 4635) [ClassicSimilarity], result of:
              0.09029783 = score(doc=4635,freq=2.0), product of:
                0.16670525 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047605187 = queryNorm
                0.5416616 = fieldWeight in 4635, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4635)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    26.11.2005 18:39:22
  19. Martins, S. de Castro: Modelo conceitual de ecossistema semântico de informações corporativas para aplicação em objetos multimídia (2019) 0.01
    0.01028779 = product of:
      0.04115116 = sum of:
        0.04115116 = product of:
          0.08230232 = sum of:
            0.08230232 = weight(_text_:model in 117) [ClassicSimilarity], result of:
              0.08230232 = score(doc=117,freq=14.0), product of:
                0.1830527 = queryWeight, product of:
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.047605187 = queryNorm
                0.44960997 = fieldWeight in 117, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  3.845226 = idf(docFreq=2569, maxDocs=44218)
                  0.03125 = fieldNorm(doc=117)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    Information management in corporate environments is a growing problem as companies' information assets grow and their need to use them in their operations. Several management models have been practiced with application on the most diverse fronts, practices that integrate the so-called Enterprise Content Management. This study proposes a conceptual model of semantic corporate information ecosystem, based on the Universal Document Model proposed by Dagobert Soergel. It focuses on unstructured information objects, especially multimedia, increasingly used in corporate environments, adding semantics and expanding their recovery potential in the composition and reuse of dynamic documents on demand. The proposed model considers stable elements in the organizational environment, such as actors, processes, business metadata and information objects, as well as some basic infrastructures of the corporate information environment. The main objective is to establish a conceptual model that adds semantic intelligence to information assets, leveraging pre-existing infrastructure in organizations, integrating and relating objects to other objects, actors and business processes. The approach methodology considered the state of the art of Information Organization, Representation and Retrieval, Organizational Content Management and Semantic Web technologies, in the scientific literature, as bases for the establishment of an integrative conceptual model. Therefore, the research will be qualitative and exploratory. The predicted steps of the model are: Environment, Data Type and Source Definition, Data Distillation, Metadata Enrichment, and Storage. As a result, in theoretical terms the extended model allows to process heterogeneous and unstructured data according to the established cut-outs and through the processes listed above, allowing value creation in the composition of dynamic information objects, with semantic aggregations to metadata.
  20. Stanz, G.: Medienarchive: Analyse einer unterschätzten Ressource : Archivierung, Dokumentation, und Informationsvermittlung in Medien bei besonderer Berücksichtigung von Pressearchiven (1994) 0.01
    0.009674768 = product of:
      0.03869907 = sum of:
        0.03869907 = product of:
          0.07739814 = sum of:
            0.07739814 = weight(_text_:22 in 9) [ClassicSimilarity], result of:
              0.07739814 = score(doc=9,freq=2.0), product of:
                0.16670525 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047605187 = queryNorm
                0.46428138 = fieldWeight in 9, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=9)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    22. 2.1997 19:50:29

Languages

  • d 29
  • e 21
  • f 1
  • hu 1
  • pt 1
  • More… Less…