Search (58 results, page 1 of 3)

  • × type_ss:"x"
  • × year_i:[2010 TO 2020}
  1. Verwer, K.: Freiheit und Verantwortung bei Hans Jonas (2011) 0.04
    0.0400367 = product of:
      0.1601468 = sum of:
        0.1601468 = product of:
          0.4804404 = sum of:
            0.4804404 = weight(_text_:3a in 973) [ClassicSimilarity], result of:
              0.4804404 = score(doc=973,freq=2.0), product of:
                0.42742437 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.050415643 = queryNorm
                1.1240361 = fieldWeight in 973, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.09375 = fieldNorm(doc=973)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Content
    Vgl.: http%3A%2F%2Fcreativechoice.org%2Fdoc%2FHansJonas.pdf&usg=AOvVaw1TM3teaYKgABL5H9yoIifA&opi=89978449.
  2. Farazi, M.: Faceted lightweight ontologies : a formalization and some experiments (2010) 0.04
    0.039432906 = product of:
      0.07886581 = sum of:
        0.06672784 = product of:
          0.20018351 = sum of:
            0.20018351 = weight(_text_:3a in 4997) [ClassicSimilarity], result of:
              0.20018351 = score(doc=4997,freq=2.0), product of:
                0.42742437 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.050415643 = queryNorm
                0.46834838 = fieldWeight in 4997, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4997)
          0.33333334 = coord(1/3)
        0.01213797 = weight(_text_:information in 4997) [ClassicSimilarity], result of:
          0.01213797 = score(doc=4997,freq=4.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.13714671 = fieldWeight in 4997, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4997)
      0.5 = coord(2/4)
    
    Content
    PhD Dissertation at International Doctorate School in Information and Communication Technology. Vgl.: https%3A%2F%2Fcore.ac.uk%2Fdownload%2Fpdf%2F150083013.pdf&usg=AOvVaw2n-qisNagpyT0lli_6QbAQ.
    Imprint
    Trento : University / Department of information engineering and computer science
  3. Xiong, C.: Knowledge based text representations for information retrieval (2016) 0.04
    0.03640151 = product of:
      0.07280302 = sum of:
        0.05338227 = product of:
          0.1601468 = sum of:
            0.1601468 = weight(_text_:3a in 5820) [ClassicSimilarity], result of:
              0.1601468 = score(doc=5820,freq=2.0), product of:
                0.42742437 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.050415643 = queryNorm
                0.3746787 = fieldWeight in 5820, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5820)
          0.33333334 = coord(1/3)
        0.019420752 = weight(_text_:information in 5820) [ClassicSimilarity], result of:
          0.019420752 = score(doc=5820,freq=16.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.21943474 = fieldWeight in 5820, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
      0.5 = coord(2/4)
    
    Abstract
    The successes of information retrieval (IR) in recent decades were built upon bag-of-words representations. Effective as it is, bag-of-words is only a shallow text understanding; there is a limited amount of information for document ranking in the word space. This dissertation goes beyond words and builds knowledge based text representations, which embed the external and carefully curated information from knowledge bases, and provide richer and structured evidence for more advanced information retrieval systems. This thesis research first builds query representations with entities associated with the query. Entities' descriptions are used by query expansion techniques that enrich the query with explanation terms. Then we present a general framework that represents a query with entities that appear in the query, are retrieved by the query, or frequently show up in the top retrieved documents. A latent space model is developed to jointly learn the connections from query to entities and the ranking of documents, modeling the external evidence from knowledge bases and internal ranking features cooperatively. To further improve the quality of relevant entities, a defining factor of our query representations, we introduce learning to rank to entity search and retrieve better entities from knowledge bases. In the document representation part, this thesis research also moves one step forward with a bag-of-entities model, in which documents are represented by their automatic entity annotations, and the ranking is performed in the entity space.
    This proposal includes plans to improve the quality of relevant entities with a co-learning framework that learns from both entity labels and document labels. We also plan to develop a hybrid ranking system that combines word based and entity based representations together with their uncertainties considered. At last, we plan to enrich the text representations with connections between entities. We propose several ways to infer entity graph representations for texts, and to rank documents using their structure representations. This dissertation overcomes the limitation of word based representations with external and carefully curated information from knowledge bases. We believe this thesis research is a solid start towards the new generation of intelligent, semantic, and structured information retrieval.
    Content
    Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Language and Information Technologies. Vgl.: https%3A%2F%2Fwww.cs.cmu.edu%2F~cx%2Fpapers%2Fknowledge_based_text_representation.pdf&usg=AOvVaw0SaTSvhWLTh__Uz_HtOtl3.
  4. Thornton, K: Powerful structure : inspecting infrastructures of information organization in Wikimedia Foundation projects (2016) 0.03
    0.026535526 = product of:
      0.05307105 = sum of:
        0.02303018 = weight(_text_:information in 3288) [ClassicSimilarity], result of:
          0.02303018 = score(doc=3288,freq=10.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.2602176 = fieldWeight in 3288, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3288)
        0.030040871 = product of:
          0.060081743 = sum of:
            0.060081743 = weight(_text_:organization in 3288) [ClassicSimilarity], result of:
              0.060081743 = score(doc=3288,freq=4.0), product of:
                0.17974974 = queryWeight, product of:
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.050415643 = queryNorm
                0.33425218 = fieldWeight in 3288, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3288)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    This dissertation investigates the social and technological factors of collaboratively organizing information in commons-based peer production systems. To do so, it analyzes the diverse strategies that members of Wikimedia Foundation (WMF) project communities use to organize information. Key findings from this dissertation show that conceptual structures of information organization are encoded into the infrastructure of WMF projects. The fact that WMF projects are commons-based peer production systems means that we can inspect the code that enables these systems, but a specific type of technical literacy is required to do so. I use three methods in this dissertation. I conduct a qualitative content analysis of the discussions surrounding the design, implementation and evaluation of the category system; a quantitative analysis using descriptive statistics of patterns of editing among editors who contributed to the code of templates for information boxes; and a close reading of the infrastructure used to create the category system, the infobox templates, and the knowledge base of structured data.
  5. Ziemba, L.: Information retrieval with concept discovery in digital collections for agriculture and natural resources (2011) 0.02
    0.017380109 = product of:
      0.034760218 = sum of:
        0.020598818 = weight(_text_:information in 4728) [ClassicSimilarity], result of:
          0.020598818 = score(doc=4728,freq=18.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.23274568 = fieldWeight in 4728, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=4728)
        0.014161401 = product of:
          0.028322803 = sum of:
            0.028322803 = weight(_text_:organization in 4728) [ClassicSimilarity], result of:
              0.028322803 = score(doc=4728,freq=2.0), product of:
                0.17974974 = queryWeight, product of:
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.050415643 = queryNorm
                0.15756798 = fieldWeight in 4728, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4728)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    The amount and complexity of information available in a digital form is already huge and new information is being produced every day. Retrieving information relevant to address a particular need becomes a significant issue. This work utilizes knowledge organization systems (KOS), such as thesauri and ontologies and applies information extraction (IE) and computational linguistics (CL) techniques to organize, manage and retrieve information stored in digital collections in the agricultural domain. Two real world applications of the approach have been developed and are available and actively used by the public. An ontology is used to manage the Water Conservation Digital Library holding a dynamic collection of various types of digital resources in the domain of urban water conservation in Florida, USA. The ontology based back-end powers a fully operational web interface, available at http://library.conservefloridawater.org. The system has demonstrated numerous benefits of the ontology application, including accurate retrieval of resources, information sharing and reuse, and has proved to effectively facilitate information management. The major difficulty encountered with the approach is that large and dynamic number of concepts makes it difficult to keep the ontology consistent and to accurately catalog resources manually. To address the aforementioned issues, a combination of IE and CL techniques, such as Vector Space Model and probabilistic parsing, with the use of Agricultural Thesaurus were adapted to automatically extract concepts important for each of the texts in the Best Management Practices (BMP) Publication Library--a collection of documents in the domain of agricultural BMPs in Florida available at http://lyra.ifas.ufl.edu/LIB. A new approach of domain-specific concept discovery with the use of Internet search engine was developed. Initial evaluation of the results indicates significant improvement in precision of information extraction. The approach presented in this work focuses on problems unique to agriculture and natural resources domain, such as domain specific concepts and vocabularies, but should be applicable to any collection of texts in digital format. It may be of potential interest for anyone who needs to effectively manage a collection of digital resources.
  6. Martins, S. de Castro: Modelo conceitual de ecossistema semântico de informações corporativas para aplicação em objetos multimídia (2019) 0.02
    0.017380109 = product of:
      0.034760218 = sum of:
        0.020598818 = weight(_text_:information in 117) [ClassicSimilarity], result of:
          0.020598818 = score(doc=117,freq=18.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.23274568 = fieldWeight in 117, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=117)
        0.014161401 = product of:
          0.028322803 = sum of:
            0.028322803 = weight(_text_:organization in 117) [ClassicSimilarity], result of:
              0.028322803 = score(doc=117,freq=2.0), product of:
                0.17974974 = queryWeight, product of:
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.050415643 = queryNorm
                0.15756798 = fieldWeight in 117, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.03125 = fieldNorm(doc=117)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Information management in corporate environments is a growing problem as companies' information assets grow and their need to use them in their operations. Several management models have been practiced with application on the most diverse fronts, practices that integrate the so-called Enterprise Content Management. This study proposes a conceptual model of semantic corporate information ecosystem, based on the Universal Document Model proposed by Dagobert Soergel. It focuses on unstructured information objects, especially multimedia, increasingly used in corporate environments, adding semantics and expanding their recovery potential in the composition and reuse of dynamic documents on demand. The proposed model considers stable elements in the organizational environment, such as actors, processes, business metadata and information objects, as well as some basic infrastructures of the corporate information environment. The main objective is to establish a conceptual model that adds semantic intelligence to information assets, leveraging pre-existing infrastructure in organizations, integrating and relating objects to other objects, actors and business processes. The approach methodology considered the state of the art of Information Organization, Representation and Retrieval, Organizational Content Management and Semantic Web technologies, in the scientific literature, as bases for the establishment of an integrative conceptual model. Therefore, the research will be qualitative and exploratory. The predicted steps of the model are: Environment, Data Type and Source Definition, Data Distillation, Metadata Enrichment, and Storage. As a result, in theoretical terms the extended model allows to process heterogeneous and unstructured data according to the established cut-outs and through the processes listed above, allowing value creation in the composition of dynamic information objects, with semantic aggregations to metadata.
  7. Shala, E.: ¬Die Autonomie des Menschen und der Maschine : gegenwärtige Definitionen von Autonomie zwischen philosophischem Hintergrund und technologischer Umsetzbarkeit (2014) 0.02
    0.01668196 = product of:
      0.06672784 = sum of:
        0.06672784 = product of:
          0.20018351 = sum of:
            0.20018351 = weight(_text_:3a in 4388) [ClassicSimilarity], result of:
              0.20018351 = score(doc=4388,freq=2.0), product of:
                0.42742437 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.050415643 = queryNorm
                0.46834838 = fieldWeight in 4388, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4388)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Footnote
    Vgl. unter: https://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=2&cad=rja&uact=8&ved=2ahUKEwizweHljdbcAhVS16QKHXcFD9QQFjABegQICRAB&url=https%3A%2F%2Fwww.researchgate.net%2Fpublication%2F271200105_Die_Autonomie_des_Menschen_und_der_Maschine_-_gegenwartige_Definitionen_von_Autonomie_zwischen_philosophischem_Hintergrund_und_technologischer_Umsetzbarkeit_Redigierte_Version_der_Magisterarbeit_Karls&usg=AOvVaw06orrdJmFF2xbCCp_hL26q.
  8. Piros, A.: Az ETO-jelzetek automatikus interpretálásának és elemzésének kérdései (2018) 0.02
    0.01668196 = product of:
      0.06672784 = sum of:
        0.06672784 = product of:
          0.20018351 = sum of:
            0.20018351 = weight(_text_:3a in 855) [ClassicSimilarity], result of:
              0.20018351 = score(doc=855,freq=2.0), product of:
                0.42742437 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.050415643 = queryNorm
                0.46834838 = fieldWeight in 855, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=855)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Content
    Vgl. auch: New automatic interpreter for complex UDC numbers. Unter: <https%3A%2F%2Fudcc.org%2Ffiles%2FAttilaPiros_EC_36-37_2014-2015.pdf&usg=AOvVaw3kc9CwDDCWP7aArpfjrs5b>
  9. Mattmann, B.: ¬Die Möglichkeiten von RDA bei der Erschliessung historischer Sondermaterialien : Analyse der Beschreibungs- und Erschließungsmöglicjkeiten von Resource Description and Access für die historischen Sondermaterialien Brief und Fotografie (2014) 0.02
    0.016597321 = product of:
      0.066389285 = sum of:
        0.066389285 = weight(_text_:standards in 1651) [ClassicSimilarity], result of:
          0.066389285 = score(doc=1651,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.29545712 = fieldWeight in 1651, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.046875 = fieldNorm(doc=1651)
      0.25 = coord(1/4)
    
    Abstract
    Der neue Erschliessungsstandard RDA beschäftigt aktuell unzählige Arbeitsgruppen und Einzelpersonen aus verschiedenen Ländern. Dabei geht es vor allem um die Anpassung einzelner Regeln und Konzepte an lokale Gegebenheiten und spezifische Medientypen. Das erklärte Ziel, den Standard auch für nicht-bibliothekarische Communities attraktiv zu machen, stand jedoch bislang im Hintergrund. Es ist unklar, ob und wie sich RDA beispielsweise in Archiven und Sondersammlungen anwenden lässt. Diese Bachelorarbeit widmet sich dieser Unsicherheit und untersucht die Möglichkeiten der Katalogisierung von ausgewählten archivalischen Ressourcen. Mittels einer Literaturanalyse werden die von Nutzern und Erschliessenden als essenziell erachteten Merkmale von Briefen und Fotografien erhoben und auf die Beschreibungsmöglichkeiten mit RDA hin untersucht. Dabei zeigt sich, dass der Standard grundsätzlich aufgeschlossen gegenüber archivalischen und historischen Ressourcen ist und sich lediglich vereinzelte Anpassungen von Begriffslisten und An- wendungsregeln empfehlen. Entscheidende Fragestellungen, wie die Abbildung von Kontextualität und Hierarchien, müssen jedoch noch geklärt werden, um die Attraktivität des Standards für nicht-bibliothekarische Communities weiter zu verbessern.
  10. Kiren, T.: ¬A clustering based indexing technique of modularized ontologies for information retrieval (2017) 0.02
    0.016541 = product of:
      0.033082 = sum of:
        0.019420752 = weight(_text_:information in 4399) [ClassicSimilarity], result of:
          0.019420752 = score(doc=4399,freq=16.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.21943474 = fieldWeight in 4399, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=4399)
        0.013661247 = product of:
          0.027322493 = sum of:
            0.027322493 = weight(_text_:22 in 4399) [ClassicSimilarity], result of:
              0.027322493 = score(doc=4399,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.15476047 = fieldWeight in 4399, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4399)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Indexing plays a vital role in Information Retrieval. With the availability of huge volume of information, it has become necessary to index the information in such a way to make easier for the end users to find the information they want efficiently and accurately. Keyword-based indexing uses words as indexing terms. It is not capable of capturing the implicit relation among terms or the semantics of the words in the document. To eliminate this limitation, ontology-based indexing came into existence, which allows semantic based indexing to solve complex and indirect user queries. Ontologies are used for document indexing which allows semantic based information retrieval. Existing ontologies or the ones constructed from scratch are used presently for indexing. Constructing ontologies from scratch is a labor-intensive task and requires extensive domain knowledge whereas use of an existing ontology may leave some important concepts in documents un-annotated. Using multiple ontologies can overcome the problem of missing out concepts to a great extent, but it is difficult to manage (changes in ontologies over time by their developers) multiple ontologies and ontology heterogeneity also arises due to ontologies constructed by different ontology developers. One possible solution to managing multiple ontologies and build from scratch is to use modular ontologies for indexing.
    Modular ontologies are built in modular manner by combining modules from multiple relevant ontologies. Ontology heterogeneity also arises during modular ontology construction because multiple ontologies are being dealt with, during this process. Ontologies need to be aligned before using them for modular ontology construction. The existing approaches for ontology alignment compare all the concepts of each ontology to be aligned, hence not optimized in terms of time and search space utilization. A new indexing technique is proposed based on modular ontology. An efficient ontology alignment technique is proposed to solve the heterogeneity problem during the construction of modular ontology. Results are satisfactory as Precision and Recall are improved by (8%) and (10%) respectively. The value of Pearsons Correlation Coefficient for degree of similarity, time, search space requirement, precision and recall are close to 1 which shows that the results are significant. Further research can be carried out for using modular ontology based indexing technique for Multimedia Information Retrieval and Bio-Medical information retrieval.
    Date
    20. 1.2015 18:30:22
  11. Huo, W.: Automatic multi-word term extraction and its application to Web-page summarization (2012) 0.02
    0.015395639 = product of:
      0.030791279 = sum of:
        0.01029941 = weight(_text_:information in 563) [ClassicSimilarity], result of:
          0.01029941 = score(doc=563,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.116372846 = fieldWeight in 563, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=563)
        0.02049187 = product of:
          0.04098374 = sum of:
            0.04098374 = weight(_text_:22 in 563) [ClassicSimilarity], result of:
              0.04098374 = score(doc=563,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.23214069 = fieldWeight in 563, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=563)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    In this thesis we propose three new word association measures for multi-word term extraction. We combine these association measures with LocalMaxs algorithm in our extraction model and compare the results of different multi-word term extraction methods. Our approach is language and domain independent and requires no training data. It can be applied to such tasks as text summarization, information retrieval, and document classification. We further explore the potential of using multi-word terms as an effective representation for general web-page summarization. We extract multi-word terms from human written summaries in a large collection of web-pages, and generate the summaries by aligning document words with these multi-word terms. Our system applies machine translation technology to learn the aligning process from a training set and focuses on selecting high quality multi-word terms from human written summaries to generate suitable results for web-page summarization.
    Date
    10. 1.2013 19:22:47
  12. Bös, K.: Aspektorientierte Inhaltserschließung von Romanen und Bildern : ein Vergleich der Ansätze von Annelise Mark Pejtersen und Sara Shatford (2012) 0.01
    0.0138311 = product of:
      0.0553244 = sum of:
        0.0553244 = weight(_text_:standards in 400) [ClassicSimilarity], result of:
          0.0553244 = score(doc=400,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.24621427 = fieldWeight in 400, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.0390625 = fieldNorm(doc=400)
      0.25 = coord(1/4)
    
    Abstract
    Für die inhaltliche Erschließung von Sach- und Fachliteratur stehen heutzutage etablierte Verfahren und Standards zur Verfügung. Anders verhält es sich dagegen mit der Erschließung von Schöner Literatur und Bildern. Beide Medien sind sehr verschieden und haben doch eines gemeinsam. Sie lassen sich mit den Regeln für Sach- und Fachliteratur nicht zufriedenstellend inhaltlich erschließen. Dieses Problem erkannten in den 1970er und 80er Jahren beide Autoren, deren Methoden ich hier verglichen habe. Annelise Mark Pejtersen bemühte sich um eine Lösung für die Schöne Literatur und wählte dabei einen empirischen Ansatz. Sara Shatford versuchte durch theoretische Überlegungen eine Lösung für Bilder zu erarbeiten. Der empirische wie der theoretische Ansatz führten zu Methoden, die das jeweilige Medium unter verschiedenen Aspekten betrachten. Diese Aspekten basieren in beiden Fällen auf denselben Fragen. Dennoch unterscheiden sie sich stark voneinander sowohl im Hinblick auf die Inhalte, die sie aufnehmen können, als auch hinsichtlich ihrer Struktur. Eine Anwendung einer der Methoden auf das jeweils andere Medium erscheint daher nicht sinnvoll. In dieser Arbeit werden die Methoden von Pejtersen und Shatford zunächst einzeln erläutert. Im Anschluss werden die Aspekte beider Methoden vergleichend gegenübergestellt. Dazu werden ausgewählte Beispiele mit beiden Methoden erschlossen. Abschließend wird geprüft, ob die wechselseitige Erschließung, wie sie im Vergleich angewendet wurde, in der Praxis sinnvoll ist und ob es Medien gibt, deren Erschließung mit beiden Methoden interessant wäre.
  13. Fischer, M.: Sacherschliessung - quo vadis? : Die Neuausrichtung der Sacherschliessung im deutschsprachigen Raum (2015) 0.01
    0.0138311 = product of:
      0.0553244 = sum of:
        0.0553244 = weight(_text_:standards in 2029) [ClassicSimilarity], result of:
          0.0553244 = score(doc=2029,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.24621427 = fieldWeight in 2029, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2029)
      0.25 = coord(1/4)
    
    Abstract
    Informationen werden heute von den meisten Menschen vor allem im World Wide Web gesucht. Bibliothekskataloge und damit die in den wissenschaftlichen Bibliotheken gepflegte intellektuelle Sacherschliessung konkurrieren dabei mit einfach und intuitiv zu benutzenden Suchmaschinen. Die Anforderungen an die thematische Recherche hat sich in den letzten Jahren durch die rasante Entwicklung der Informationstechnologie grundlegend verändert. Darüber hinaus sehen sich die Bibliotheken heute mit dem Problem konfrontiert, dass die zunehmende Flut an elektronischen Publikationen bei gleichzeitig abnehmenden Ressourcen intellektuell nicht mehr bewältigt werden kann. Vor diesem Hintergrund hat die Expertengruppe Sacherschliessung - eine Arbeitsgruppe innerhalb der Arbeitsstelle für Standardisierung der Deutschen Nationalbibliothek (DNB), in welcher Vertreterinnen und Vertreter der deutschsprachigen Bibliotheksverbünde repräsentiert sind - 2013 damit begonnen, sich mit der Neuausrichtung der verbalen Sacherschliessung zu befassen. Bei der aktuellen Überarbeitung der Regeln für den Schlagwortkatalog (RSWK) sollen die verbale und klassifikatorische Sacherschliessung, ebenso wie die intellektuelle und automatische Indexierung in einem Zusammenhang betrachtet werden. Neben der neuen Suchmaschinentechnologie und den automatischen Indexierungsmethoden gewinnt dabei vor allem die Vernetzung der Bibliothekskataloge mit anderen Ressourcen im World Wide Web immer mehr an Bedeutung. Ausgehend von einer Analyse der grundlegenden Prinzipien der international verbreiteten Normen und Standards (FRBR, FRSAD und RDA) beschäftige ich mich in meiner Masterarbeit mit der in der Expertengruppe Sacherschliessung geführten Debatte über die aktuelle Überarbeitung der RSWK. Dabei stellt sich insbesondere die Frage, welche Auswirkungen die rasante Entwicklung der Informationstechnologie auf die zukünftige Neuausrichtung der intellektuellen Sacherschliessung haben wird? Welche Rolle spielen in Zukunft die Suchmaschinen und Discovery Systeme, die automatischen Indexierungsverfahren und das Semantic Web bzw. Linked Open Data bei der inhaltlichen Erschliessung von bibliografischen Ressourcen?
  14. Geisriegler, E.: Enriching electronic texts with semantic metadata : a use case for the historical Newspaper Collection ANNO (Austrian Newspapers Online) of the Austrian National Libraryhek (2012) 0.01
    0.0128297005 = product of:
      0.025659401 = sum of:
        0.008582841 = weight(_text_:information in 595) [ClassicSimilarity], result of:
          0.008582841 = score(doc=595,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.09697737 = fieldWeight in 595, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=595)
        0.01707656 = product of:
          0.03415312 = sum of:
            0.03415312 = weight(_text_:22 in 595) [ClassicSimilarity], result of:
              0.03415312 = score(doc=595,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.19345059 = fieldWeight in 595, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=595)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Date
    3. 2.2013 18:00:22
    Footnote
    Wien, Univ., Lehrgang Library and Information Studies, Master-Thesis, 2012.
  15. Petit, A.L.B.: Ordnungssysteme im Records Management (2014) 0.01
    0.0110648805 = product of:
      0.044259522 = sum of:
        0.044259522 = weight(_text_:standards in 2305) [ClassicSimilarity], result of:
          0.044259522 = score(doc=2305,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.19697142 = fieldWeight in 2305, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.03125 = fieldNorm(doc=2305)
      0.25 = coord(1/4)
    
    Abstract
    Immer dann, wenn ein Paradigma in Frage gestellt werden soll, ist es unabdingbar, sich mit dem Grund seiner Existenz auseinander zu setzen. So sollen nach eingehenden Begriffsklärungen des Records Management und der Ordnungstheorie, die grundlegend für diese Analyse sind, in einem nächsten Schritt die unterschiedlichen Ordnungselemente im Records Management untersucht und ihre Rolle im Gesamtkontext der Disziplin eruiert werden. Dies wird im weiteren Verlauf der Arbeit insbesondere durch eine Analyse diverser Standards sowie der Norm ISO-15489 und deren Implikationen an Ordnungssysteme gewährleistet. Nachdem auf diese Weise die Ordnungssysteme in ihrer Rolle als Informationsverwaltungswerkzeug positioniert wurden, soll im nächsten Teil der Arbeit der Fokus auf das Retrieval und die Interaktion mit dem Nutzer gelegt werden. In diesem Zusammenhang wird das Prinzip Hierarchie, welches den Aktenplan strukturell definiert, in einem ersten Schritt mit den kognitiven Strukturen des Menschen abgeglichen, um hier etwaige Diskrepanzen festzustellen oder ihn zu legitimieren. Daraufhin soll in einem Exkurs in die Informationsverhaltensforschung und empirische Nutzerstudien von Records Management-Systemen versucht werden, die bevorzugten Suchstrategien in Bezug auf Browsing oder Searching herauszustellen. Auf der Grundlage dieser Erkenntnisse sollen im letzten Teil der Arbeit die Konsequenzen für die Ordnungssysteme im Records Management dargestellt sowie die Verbindung zu weiteren technologischen Entwicklungen angerissen werden. Im Fazit werden die Ergebnisse zusammenfassend rekapituliert und abschließend bewertet.
  16. Lamparter, A.: Kompetenzprofil von Information Professionals in Unternehmen (2015) 0.01
    0.00849658 = product of:
      0.03398632 = sum of:
        0.03398632 = weight(_text_:information in 769) [ClassicSimilarity], result of:
          0.03398632 = score(doc=769,freq=16.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.3840108 = fieldWeight in 769, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=769)
      0.25 = coord(1/4)
    
    Abstract
    Information Professionals sind in Unternehmen für den professionellen und strategischen Umgang mit Informationen verantwortlich. Da es keine allgemeingültige Definition für diese Berufsgruppe gibt, wird in der Masterarbeit eine Begriffsbestimmung unternommen. Mit Hilfe dreier Methoden, einer Auswertung von relevanter Fachliteratur, die Untersuchung von einschlägigen Stellenausschreibungen und das Führen von Experteninterviews, wird ein Kompetenzprofil für Information Professionals erstellt. 16 Kompetenzen in den Bereichen Fach-, Methoden-, Sozial- und persönliche Kompetenzen geben eine Orientierung über vorhandene Fähigkeiten dieser Berufsgruppe für Personalfachleute, Vorgesetzte und Information Professionals selbst.
    Content
    Masterarbeit an der Hochschule Hannover, Fakultät III - Medien, Information und Design. Trägerin des VFI-Förderpreises 2015, Vgl.: urn:nbn:de:bsz:960-opus4-5280. http://serwiss.bib.hs-hannover.de/frontdoor/index/index/docId/528. Vgl. auch: Knoll, A. (geb. Lamparter): Kompetenzprofil von Information Professionals in Unternehmen. In: Young information professionals. 1(2016) S.1-11.
    Imprint
    Hannover : Hochschule, Fakultät III - Medien, Information und Design
  17. Glaesener, L.: Automatisches Indexieren einer informationswissenschaftlichen Datenbank mit Mehrwortgruppen (2012) 0.01
    0.0068306234 = product of:
      0.027322493 = sum of:
        0.027322493 = product of:
          0.054644987 = sum of:
            0.054644987 = weight(_text_:22 in 401) [ClassicSimilarity], result of:
              0.054644987 = score(doc=401,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.30952093 = fieldWeight in 401, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=401)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    11. 9.2012 19:43:22
  18. Schmolz, H.: Anaphora resolution and text retrieval : a lnguistic analysis of hypertexts (2015) 0.01
    0.006068985 = product of:
      0.02427594 = sum of:
        0.02427594 = weight(_text_:information in 1172) [ClassicSimilarity], result of:
          0.02427594 = score(doc=1172,freq=4.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.27429342 = fieldWeight in 1172, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=1172)
      0.25 = coord(1/4)
    
    RSWK
    Englisch / Anapher <Syntax> / Hypertext / Information Retrieval / Korpus <Linguistik>
    Subject
    Englisch / Anapher <Syntax> / Hypertext / Information Retrieval / Korpus <Linguistik>
  19. Bertram, J.: Informationen verzweifelt gesucht : Enterprise Search in österreichischen Großunternehmen (2011) 0.01
    0.006037475 = product of:
      0.0241499 = sum of:
        0.0241499 = product of:
          0.0482998 = sum of:
            0.0482998 = weight(_text_:22 in 2657) [ClassicSimilarity], result of:
              0.0482998 = score(doc=2657,freq=4.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.27358043 = fieldWeight in 2657, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2657)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    Die Arbeit geht dem Status quo der unternehmensweiten Suche in österreichischen Großunternehmen nach und beleuchtet Faktoren, die darauf Einfluss haben. Aus der Analyse des Ist-Zustands wird der Bedarf an Enterprise-Search-Software abgeleitet und es werden Rahmenbedingungen für deren erfolgreiche Einführung skizziert. Die Untersuchung stützt sich auf eine im Jahr 2009 durchgeführte Onlinebefragung von 469 österreichischen Großunternehmen (Rücklauf 22 %) und daran anschließende Leitfadeninterviews mit zwölf Teilnehmern der Onlinebefragung. Der theoretische Teil verortet die Arbeit im Kontext des Informations- und Wissensmanagements. Der Fokus liegt auf dem Ansatz der Enterprise Search, ihrer Abgrenzung gegenüber der Suche im Internet und ihrem Leistungsspektrum. Im empirischen Teil wird zunächst aufgezeigt, wie die Unternehmen ihre Informationen organisieren und welche Probleme dabei auftreten. Es folgt eine Analyse des Status quo der Informati-onssuche im Unternehmen. Abschließend werden Bekanntheit und Einsatz von Enterprise-Search-Software in der Zielgruppe untersucht sowie für die Einführung dieser Software nötige Rahmenbedingungen benannt. Defizite machen die Befragten insbesondere im Hinblick auf die übergreifende Suche im Unternehmen und die Suche nach Kompetenzträgern aus. Hier werden Lücken im Wissensmanagement offenbar. 29 % der Respondenten der Onlinebefragung geben zu-dem an, dass es in ihren Unternehmen gelegentlich bis häufig zu Fehlentscheidungen infolge defizitärer Informationslagen kommt. Enterprise-Search-Software kommt in 17 % der Unternehmen, die sich an der Onlinebefragung beteiligten, zum Einsatz. Die durch Enterprise-Search-Software bewirkten Veränderungen werden grundsätzlich posi-tiv beurteilt. Alles in allem zeigen die Ergebnisse, dass Enterprise-Search-Strategien nur Erfolg haben können, wenn man sie in umfassende Maßnahmen des Informations- und Wissensmanagements einbettet.
    Date
    22. 1.2016 20:40:31
  20. Smith, D.A.: Exploratory and faceted browsing over heterogeneous and cross-domain data sources. (2011) 0.01
    0.005149705 = product of:
      0.02059882 = sum of:
        0.02059882 = weight(_text_:information in 4839) [ClassicSimilarity], result of:
          0.02059882 = score(doc=4839,freq=8.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.23274569 = fieldWeight in 4839, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=4839)
      0.25 = coord(1/4)
    
    Abstract
    Exploration of heterogeneous data sources increases the value of information by allowing users to answer questions through exploration across multiple sources; Users can use information that has been posted across the Web to answer questions and learn about new domains. We have conducted research that lowers the interrogation time of faceted data, by combining related information from different sources. The work contributes methodologies in combining heterogenous sources, and how to deliver that data to a user interface scalably, with enough performance to support rapid interrogation of the knowledge by the user. The work also contributes how to combine linked data sources so that users can create faceted browsers that target the information facets of their needs. The work is grounded and proven in a number of experiments and test cases that study the contributions in domain research work.

Languages

  • d 35
  • e 18
  • a 1
  • f 1
  • hu 1
  • pt 1
  • More… Less…

Types