Search (258 results, page 1 of 13)

  • × type_ss:"x"
  1. Verwer, K.: Freiheit und Verantwortung bei Hans Jonas (2011) 0.04
    0.0400367 = product of:
      0.1601468 = sum of:
        0.1601468 = product of:
          0.4804404 = sum of:
            0.4804404 = weight(_text_:3a in 973) [ClassicSimilarity], result of:
              0.4804404 = score(doc=973,freq=2.0), product of:
                0.42742437 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.050415643 = queryNorm
                1.1240361 = fieldWeight in 973, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.09375 = fieldNorm(doc=973)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Content
    Vgl.: http%3A%2F%2Fcreativechoice.org%2Fdoc%2FHansJonas.pdf&usg=AOvVaw1TM3teaYKgABL5H9yoIifA&opi=89978449.
  2. Farazi, M.: Faceted lightweight ontologies : a formalization and some experiments (2010) 0.04
    0.039432906 = product of:
      0.07886581 = sum of:
        0.06672784 = product of:
          0.20018351 = sum of:
            0.20018351 = weight(_text_:3a in 4997) [ClassicSimilarity], result of:
              0.20018351 = score(doc=4997,freq=2.0), product of:
                0.42742437 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.050415643 = queryNorm
                0.46834838 = fieldWeight in 4997, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4997)
          0.33333334 = coord(1/3)
        0.01213797 = weight(_text_:information in 4997) [ClassicSimilarity], result of:
          0.01213797 = score(doc=4997,freq=4.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.13714671 = fieldWeight in 4997, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4997)
      0.5 = coord(2/4)
    
    Content
    PhD Dissertation at International Doctorate School in Information and Communication Technology. Vgl.: https%3A%2F%2Fcore.ac.uk%2Fdownload%2Fpdf%2F150083013.pdf&usg=AOvVaw2n-qisNagpyT0lli_6QbAQ.
    Imprint
    Trento : University / Department of information engineering and computer science
  3. Gabler, S.: Vergabe von DDC-Sachgruppen mittels eines Schlagwort-Thesaurus (2021) 0.04
    0.039432906 = product of:
      0.07886581 = sum of:
        0.06672784 = product of:
          0.20018351 = sum of:
            0.20018351 = weight(_text_:3a in 1000) [ClassicSimilarity], result of:
              0.20018351 = score(doc=1000,freq=2.0), product of:
                0.42742437 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.050415643 = queryNorm
                0.46834838 = fieldWeight in 1000, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1000)
          0.33333334 = coord(1/3)
        0.01213797 = weight(_text_:information in 1000) [ClassicSimilarity], result of:
          0.01213797 = score(doc=1000,freq=4.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.13714671 = fieldWeight in 1000, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1000)
      0.5 = coord(2/4)
    
    Content
    Master thesis Master of Science (Library and Information Studies) (MSc), Universität Wien. Advisor: Christoph Steiner. Vgl.: https://www.researchgate.net/publication/371680244_Vergabe_von_DDC-Sachgruppen_mittels_eines_Schlagwort-Thesaurus. DOI: 10.25365/thesis.70030. Vgl. dazu die Präsentation unter: https://www.google.com/url?sa=i&rct=j&q=&esrc=s&source=web&cd=&ved=0CAIQw7AJahcKEwjwoZzzytz_AhUAAAAAHQAAAAAQAg&url=https%3A%2F%2Fwiki.dnb.de%2Fdownload%2Fattachments%2F252121510%2FDA3%2520Workshop-Gabler.pdf%3Fversion%3D1%26modificationDate%3D1671093170000%26api%3Dv2&psig=AOvVaw0szwENK1or3HevgvIDOfjx&ust=1687719410889597&opi=89978449.
    Imprint
    Wien / Library and Information Studies : Universität
  4. König, E.: Nationale und internationale Standards und Empfehlungen für den Informationsdienst : Bestandsaufnahme und Vergleich im Hinblick auf künftige Planungen und Konzeptionen für Öffentliche Bibliotheken in der Bundesrepublik Deutschland (1979) 0.04
    0.038727082 = product of:
      0.15490833 = sum of:
        0.15490833 = weight(_text_:standards in 6836) [ClassicSimilarity], result of:
          0.15490833 = score(doc=6836,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.68939996 = fieldWeight in 6836, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.109375 = fieldNorm(doc=6836)
      0.25 = coord(1/4)
    
  5. Stojanovic, N.: Ontology-based Information Retrieval : methods and tools for cooperative query answering (2005) 0.04
    0.036990546 = product of:
      0.07398109 = sum of:
        0.05338227 = product of:
          0.1601468 = sum of:
            0.1601468 = weight(_text_:3a in 701) [ClassicSimilarity], result of:
              0.1601468 = score(doc=701,freq=2.0), product of:
                0.42742437 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.050415643 = queryNorm
                0.3746787 = fieldWeight in 701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
          0.33333334 = coord(1/3)
        0.020598818 = weight(_text_:information in 701) [ClassicSimilarity], result of:
          0.020598818 = score(doc=701,freq=18.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.23274568 = fieldWeight in 701, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
      0.5 = coord(2/4)
    
    Abstract
    By the explosion of possibilities for a ubiquitous content production, the information overload problem reaches the level of complexity which cannot be managed by traditional modelling approaches anymore. Due to their pure syntactical nature traditional information retrieval approaches did not succeed in treating content itself (i.e. its meaning, and not its representation). This leads to a very low usefulness of the results of a retrieval process for a user's task at hand. In the last ten years ontologies have been emerged from an interesting conceptualisation paradigm to a very promising (semantic) modelling technology, especially in the context of the Semantic Web. From the information retrieval point of view, ontologies enable a machine-understandable form of content description, such that the retrieval process can be driven by the meaning of the content. However, the very ambiguous nature of the retrieval process in which a user, due to the unfamiliarity with the underlying repository and/or query syntax, just approximates his information need in a query, implies a necessity to include the user in the retrieval process more actively in order to close the gap between the meaning of the content and the meaning of a user's query (i.e. his information need). This thesis lays foundation for such an ontology-based interactive retrieval process, in which the retrieval system interacts with a user in order to conceptually interpret the meaning of his query, whereas the underlying domain ontology drives the conceptualisation process. In that way the retrieval process evolves from a query evaluation process into a highly interactive cooperation between a user and the retrieval system, in which the system tries to anticipate the user's information need and to deliver the relevant content proactively. Moreover, the notion of content relevance for a user's query evolves from a content dependent artefact to the multidimensional context-dependent structure, strongly influenced by the user's preferences. This cooperation process is realized as the so-called Librarian Agent Query Refinement Process. In order to clarify the impact of an ontology on the retrieval process (regarding its complexity and quality), a set of methods and tools for different levels of content and query formalisation is developed, ranging from pure ontology-based inferencing to keyword-based querying in which semantics automatically emerges from the results. Our evaluation studies have shown that the possibilities to conceptualize a user's information need in the right manner and to interpret the retrieval results accordingly are key issues for realizing much more meaningful information retrieval systems.
    Content
    Vgl.: http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F1627&ei=tAtYUYrBNoHKtQb3l4GYBw&usg=AFQjCNHeaxKkKU3-u54LWxMNYGXaaDLCGw&sig2=8WykXWQoDKjDSdGtAakH2Q&bvm=bv.44442042,d.Yms.
  6. Xiong, C.: Knowledge based text representations for information retrieval (2016) 0.04
    0.03640151 = product of:
      0.07280302 = sum of:
        0.05338227 = product of:
          0.1601468 = sum of:
            0.1601468 = weight(_text_:3a in 5820) [ClassicSimilarity], result of:
              0.1601468 = score(doc=5820,freq=2.0), product of:
                0.42742437 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.050415643 = queryNorm
                0.3746787 = fieldWeight in 5820, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5820)
          0.33333334 = coord(1/3)
        0.019420752 = weight(_text_:information in 5820) [ClassicSimilarity], result of:
          0.019420752 = score(doc=5820,freq=16.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.21943474 = fieldWeight in 5820, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
      0.5 = coord(2/4)
    
    Abstract
    The successes of information retrieval (IR) in recent decades were built upon bag-of-words representations. Effective as it is, bag-of-words is only a shallow text understanding; there is a limited amount of information for document ranking in the word space. This dissertation goes beyond words and builds knowledge based text representations, which embed the external and carefully curated information from knowledge bases, and provide richer and structured evidence for more advanced information retrieval systems. This thesis research first builds query representations with entities associated with the query. Entities' descriptions are used by query expansion techniques that enrich the query with explanation terms. Then we present a general framework that represents a query with entities that appear in the query, are retrieved by the query, or frequently show up in the top retrieved documents. A latent space model is developed to jointly learn the connections from query to entities and the ranking of documents, modeling the external evidence from knowledge bases and internal ranking features cooperatively. To further improve the quality of relevant entities, a defining factor of our query representations, we introduce learning to rank to entity search and retrieve better entities from knowledge bases. In the document representation part, this thesis research also moves one step forward with a bag-of-entities model, in which documents are represented by their automatic entity annotations, and the ranking is performed in the entity space.
    This proposal includes plans to improve the quality of relevant entities with a co-learning framework that learns from both entity labels and document labels. We also plan to develop a hybrid ranking system that combines word based and entity based representations together with their uncertainties considered. At last, we plan to enrich the text representations with connections between entities. We propose several ways to infer entity graph representations for texts, and to rank documents using their structure representations. This dissertation overcomes the limitation of word based representations with external and carefully curated information from knowledge bases. We believe this thesis research is a solid start towards the new generation of intelligent, semantic, and structured information retrieval.
    Content
    Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Language and Information Technologies. Vgl.: https%3A%2F%2Fwww.cs.cmu.edu%2F~cx%2Fpapers%2Fknowledge_based_text_representation.pdf&usg=AOvVaw0SaTSvhWLTh__Uz_HtOtl3.
  7. Sperling, R.: Anlage von Literaturreferenzen für Onlineressourcen auf einer virtuellen Lernplattform (2004) 0.04
    0.03592316 = product of:
      0.07184632 = sum of:
        0.024031956 = weight(_text_:information in 4635) [ClassicSimilarity], result of:
          0.024031956 = score(doc=4635,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.27153665 = fieldWeight in 4635, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=4635)
        0.047814365 = product of:
          0.09562873 = sum of:
            0.09562873 = weight(_text_:22 in 4635) [ClassicSimilarity], result of:
              0.09562873 = score(doc=4635,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.5416616 = fieldWeight in 4635, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4635)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Date
    26.11.2005 18:39:22
    Imprint
    Potsdam : Fachhochschule, Institut für Information und Dokumentation
  8. Milanesi, C.: Möglichkeiten der Kooperation im Rahmen von Subject Gateways : das Euler-Projekt im Vergleich mit weiteren europäischen Projekten (2001) 0.03
    0.030791279 = product of:
      0.061582558 = sum of:
        0.02059882 = weight(_text_:information in 4865) [ClassicSimilarity], result of:
          0.02059882 = score(doc=4865,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.23274569 = fieldWeight in 4865, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=4865)
        0.04098374 = product of:
          0.08196748 = sum of:
            0.08196748 = weight(_text_:22 in 4865) [ClassicSimilarity], result of:
              0.08196748 = score(doc=4865,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.46428138 = fieldWeight in 4865, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4865)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Date
    22. 6.2002 19:41:59
    Theme
    Information Gateway
  9. Thornton, K: Powerful structure : inspecting infrastructures of information organization in Wikimedia Foundation projects (2016) 0.03
    0.026535526 = product of:
      0.05307105 = sum of:
        0.02303018 = weight(_text_:information in 3288) [ClassicSimilarity], result of:
          0.02303018 = score(doc=3288,freq=10.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.2602176 = fieldWeight in 3288, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3288)
        0.030040871 = product of:
          0.060081743 = sum of:
            0.060081743 = weight(_text_:organization in 3288) [ClassicSimilarity], result of:
              0.060081743 = score(doc=3288,freq=4.0), product of:
                0.17974974 = queryWeight, product of:
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.050415643 = queryNorm
                0.33425218 = fieldWeight in 3288, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3288)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    This dissertation investigates the social and technological factors of collaboratively organizing information in commons-based peer production systems. To do so, it analyzes the diverse strategies that members of Wikimedia Foundation (WMF) project communities use to organize information. Key findings from this dissertation show that conceptual structures of information organization are encoded into the infrastructure of WMF projects. The fact that WMF projects are commons-based peer production systems means that we can inspect the code that enables these systems, but a specific type of technical literacy is required to do so. I use three methods in this dissertation. I conduct a qualitative content analysis of the discussions surrounding the design, implementation and evaluation of the category system; a quantitative analysis using descriptive statistics of patterns of editing among editors who contributed to the code of templates for information boxes; and a close reading of the infrastructure used to create the category system, the infobox templates, and the knowledge base of structured data.
  10. Makewita, S.M.: Investigating the generic information-seeking function of organisational decision-makers : perspectives on improving organisational information systems (2002) 0.02
    0.024595305 = product of:
      0.04919061 = sum of:
        0.03211405 = weight(_text_:information in 642) [ClassicSimilarity], result of:
          0.03211405 = score(doc=642,freq=28.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.3628561 = fieldWeight in 642, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=642)
        0.01707656 = product of:
          0.03415312 = sum of:
            0.03415312 = weight(_text_:22 in 642) [ClassicSimilarity], result of:
              0.03415312 = score(doc=642,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.19345059 = fieldWeight in 642, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=642)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    The past decade has seen the emergence of a new paradigm in the corporate world where organisations emphasised connectivity as a means of exposing decision-makers to wider resources of information within and outside the organisation. Many organisations followed the initiatives of enhancing infrastructures, manipulating cultural shifts and emphasising managerial commitment for creating pools and networks of knowledge. However, the concept of connectivity is not merely presenting people with the data, but more importantly, to create environments where people can seek information efficiently. This paradigm has therefore caused a shift in the function of information systems in organisations. They have to be now assessed in relation to how they underpin people's information-seeking activities within the context of their organisational environment. This research project used interpretative research methods to investigate the nature of people's information-seeking activities at two culturally contrasting organisations. Outcomes of this research project provide insights into phenomena associated with people's information-seeking function, and show how they depend on the organisational context that is defined partly by information systems. It suggests that information-seeking is not just searching for data. The inefficiencies inherent in both people and their environments can bring opaqueness into people's data, which they need to avoid or eliminate as part of seeking information. This seems to have made information-seeking a two-tier process consisting of a primary process of searching and interpreting data and auxiliary process of avoiding and eliminating opaqueness in data. Based on this view, this research suggests that organisational information systems operate naturally as implicit dual-mechanisms to underpin the above two-tier process, and that improvements to information systems should concern maintaining the balance in these dual-mechanisms.
    Date
    22. 7.2022 12:16:58
  11. Liebwald, D.: Evaluierung juristischer Datenbanken (2003) 0.02
    0.02361183 = product of:
      0.04722366 = sum of:
        0.00849658 = weight(_text_:information in 2490) [ClassicSimilarity], result of:
          0.00849658 = score(doc=2490,freq=4.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.0960027 = fieldWeight in 2490, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02734375 = fieldNorm(doc=2490)
        0.038727082 = weight(_text_:standards in 2490) [ClassicSimilarity], result of:
          0.038727082 = score(doc=2490,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.17234999 = fieldWeight in 2490, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.02734375 = fieldNorm(doc=2490)
      0.5 = coord(2/4)
    
    Footnote
    Rez. in Mitt. VÖB 57(2004) H.2, S.71-73 (J. Pauser):"Bei der hier zu besprechenden Arbeit handelt es sich um die Ende 2003 erfolgte Drucklegung einer juristischen Dissertation an der Universität Wien. Die Autorin zielt darauf ab, "Grundlagen, Entstehung, verschiedene Ansätze und Entwicklungstendenzen desjuristischen Information Retrieval aufzuzeigen [...], um schließlich die Qualität der wichtigsten österreichischen Rechtsdatenbanken anhand der gewonnenen Erkenntnisse messen zu können". Das gewählte Thema ist spannend und wohl für jeden Informationswissenschaftler und praktischen Juristen von Relevanz. Elektronische Datenbanken mit Rechtsinformation, seien sie nun online oder offline, revolutionieren seit geraumer Zeit die juristische Arbeit nicht nur in Österreich. Das Recherchieren mittels dieser neuen "Werkzeuge" gehört bereits standardmäßig zur Grundausbildung eines jedes Juristen. Die Kenntnis der umfassenden Möglichkeiten dieser neuen juristischen Informationsquellen beeinflusst massiv die Qualität und vor allem Schnelligkeit des juristischen Arbeitens. Vor diesem Hintergrund ist es immens wichtig, dass die juristischen Datenbanken den Bedürfnissen der Nutzer möglichst zweckmäßig entgegenkommen. Doris Liebwald definiert im ersten Teil ihrer Arbeit den Begriff "Information Retrieval" als "Repräsentation, Speicherung und Organisation von Informationen und der Zugriff auf Informationen" und versucht anschließend Bewertungskriterien für Rechtsdatenbanken aufzustellen. Hinsichtlich der Daten (Gesetze, Gerichtsentscheidungen, juristische Literatur) fordert sie "Vollständigkeit", "Aktualität" und "Authentizität"; in technischer Sicht ergeben sich "Inhaltserschließung", "Suchfunktionen/-möglichkeiten", sowie die "Benutzerfreundlichkeit der Systemoberfläche" etwa durch einfache Bedienbarkeit, Verständlichkeit, Anreicherung durch Hilfefunktionen usw. als Bewertungskriterien, Schließlich sind aus praktisch-ökonomischer Sicht noch "Kosten" und der "Support" aufgenommen.
    Der Hauptteil der Studie umfasst nun fragebogenartig strukturierte Angaben zu insgesamt 18 österreichischen Rechtsinformationsdatenbanken und zwar: Das große - aber wohl nicht zu verhindern gewesene - Manko dieses Abschnittes ist nun, dass er den Stand von Frühjahr/Sommer 2002 wiedergibt und damit zum größten Teil nicht mehr aktuell ist, denn auf dem Rechtsinformationsmarkt ergaben sich in Ietzter Zeit eine Vielzahl an Änderungen durch Verlags- und Produktverschränkungen, technische Neuerungen und Neugestaltung vieler online-Datenbanken. Der Dokumentationsteil hat damit heute fast nur mehr rechtshistorischen Wert. Beispiele: Die "Arbeitsrecht CD-Rom" wurde bereits vom Markt genommen. - "jusline-pro" dürfte auch nicht mehr existieren, zumindest erscheint im Netz an dem angegebenen URL nun schon seit geraumer Zeit ein Hin weis darauf, dass das System "momentan" (sic) nicht zur Verfügung steht. - "Orac Online" und "ARD Online" wurden mittlerweile verändert und zusammengeführt. - Der Parlamentsserver (http://www.parlinkom.gv.at) hat einen umfassenden grafischen Relaunch hinter sich gebracht, der sich leider nicht zum Vorteil des Nutzers auswirkt. Zugegeben, das alte Design wirkte billig, veraltet und unschön, aber man fand schnell und zügig die "Stenographischen Protokolle" und die "Materialien". Das neue Design fällt trotzdem weit hinter die praktikablen Standards anderer europäischer Parlamente zurück. Leider wurde hier eine Chance vertan. Die Angesprochenen rechtsrelevanten Inhalte sind zwar weiterhin und in großer Menge vorhanden, doch erfordert die Suche nun ein gewisses Maß an archivarischem Fleiß. Viele Nutzer, die sich zuvor sehr leicht getan haben, hatten nach der Umstellung der Website massive Probleme. Vielleicht hätte man hier sinnvollerweise die Publikums-Website mit den allgemeinen Daten zum Parlamentvon der Datenbank-Website mit den gesetzgebungsrelevanten Daten trennen und letztere benutzerfreundlicher aufbereiten sollen. - Die RDB hat nun ein neues Content-Portal eröffnet, das mit dem alten, hier dargestellten, kaum mehr etwas gemeinsam hat. - Im RIS werden seit Anfang 2004 die Bundesgesetzblätter authentisch kundgemacht. - Die Steuerdatenbank ist nunmehr mit der SWK-Online vereinigt. Das mag hier als Hinweis genügen. Ein kurzes zehnseitiges "Ergebnis" beschließt die Studie. Hierwurde die Chance vertan, die dargestellten Datenbanken umfassender zu vergleichen und Unterschiede, Vorteile, Nutzungsmöglichkeiten, Praxisvor- und -nach teile usw. etwa mittels eines standardisierten Tests herauszuarbeiten. Die Arbeit wird ihren Wert vor allem darin entfalten, Datenbankanbieter zu einer Verbesserung ihrer Datenbanken anzuregen. Der Hinweis, dass "die dem User zur Verfügung gestellten Hilfefunktionen und Informationen als untauglich oder unzureichend zu bewerten sind" (S. 191), sollten sich wirklich alle Datenbankanbieter auf diesem Gebiet zu Herzen nehmen."
  12. Ingold, M.: ¬Das bibliothekarische Konzept der Informationskompetenz : ein Überblick (2005) 0.02
    0.022129761 = product of:
      0.088519044 = sum of:
        0.088519044 = weight(_text_:standards in 1413) [ClassicSimilarity], result of:
          0.088519044 = score(doc=1413,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.39394283 = fieldWeight in 1413, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.0625 = fieldNorm(doc=1413)
      0.25 = coord(1/4)
    
    Abstract
    Informationskompetenz hat sich in den letzten Jahren als wichtiges Teilgebiet des Bibliothekswesens etabliert. Der vorliegende Text gibt einen einführenden Überblick über Hintergrund und Entwicklung, Terminologie, Definitionen und Inhalte, Modelle und Standards sowie Vermittlungs- und Evaluationspraxis von Informationskompetenz im bibliothekarischen Kontext. Ein Blick auf Debatten und Kritik zeigt aber auch, dass sowohl Begriff wie Konzept in bibliothekarischen Fachkreisen keineswegs unumstritten sind.
  13. Ziemba, L.: Information retrieval with concept discovery in digital collections for agriculture and natural resources (2011) 0.02
    0.017380109 = product of:
      0.034760218 = sum of:
        0.020598818 = weight(_text_:information in 4728) [ClassicSimilarity], result of:
          0.020598818 = score(doc=4728,freq=18.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.23274568 = fieldWeight in 4728, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=4728)
        0.014161401 = product of:
          0.028322803 = sum of:
            0.028322803 = weight(_text_:organization in 4728) [ClassicSimilarity], result of:
              0.028322803 = score(doc=4728,freq=2.0), product of:
                0.17974974 = queryWeight, product of:
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.050415643 = queryNorm
                0.15756798 = fieldWeight in 4728, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4728)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    The amount and complexity of information available in a digital form is already huge and new information is being produced every day. Retrieving information relevant to address a particular need becomes a significant issue. This work utilizes knowledge organization systems (KOS), such as thesauri and ontologies and applies information extraction (IE) and computational linguistics (CL) techniques to organize, manage and retrieve information stored in digital collections in the agricultural domain. Two real world applications of the approach have been developed and are available and actively used by the public. An ontology is used to manage the Water Conservation Digital Library holding a dynamic collection of various types of digital resources in the domain of urban water conservation in Florida, USA. The ontology based back-end powers a fully operational web interface, available at http://library.conservefloridawater.org. The system has demonstrated numerous benefits of the ontology application, including accurate retrieval of resources, information sharing and reuse, and has proved to effectively facilitate information management. The major difficulty encountered with the approach is that large and dynamic number of concepts makes it difficult to keep the ontology consistent and to accurately catalog resources manually. To address the aforementioned issues, a combination of IE and CL techniques, such as Vector Space Model and probabilistic parsing, with the use of Agricultural Thesaurus were adapted to automatically extract concepts important for each of the texts in the Best Management Practices (BMP) Publication Library--a collection of documents in the domain of agricultural BMPs in Florida available at http://lyra.ifas.ufl.edu/LIB. A new approach of domain-specific concept discovery with the use of Internet search engine was developed. Initial evaluation of the results indicates significant improvement in precision of information extraction. The approach presented in this work focuses on problems unique to agriculture and natural resources domain, such as domain specific concepts and vocabularies, but should be applicable to any collection of texts in digital format. It may be of potential interest for anyone who needs to effectively manage a collection of digital resources.
  14. Martins, S. de Castro: Modelo conceitual de ecossistema semântico de informações corporativas para aplicação em objetos multimídia (2019) 0.02
    0.017380109 = product of:
      0.034760218 = sum of:
        0.020598818 = weight(_text_:information in 117) [ClassicSimilarity], result of:
          0.020598818 = score(doc=117,freq=18.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.23274568 = fieldWeight in 117, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=117)
        0.014161401 = product of:
          0.028322803 = sum of:
            0.028322803 = weight(_text_:organization in 117) [ClassicSimilarity], result of:
              0.028322803 = score(doc=117,freq=2.0), product of:
                0.17974974 = queryWeight, product of:
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.050415643 = queryNorm
                0.15756798 = fieldWeight in 117, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5653565 = idf(docFreq=3399, maxDocs=44218)
                  0.03125 = fieldNorm(doc=117)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Information management in corporate environments is a growing problem as companies' information assets grow and their need to use them in their operations. Several management models have been practiced with application on the most diverse fronts, practices that integrate the so-called Enterprise Content Management. This study proposes a conceptual model of semantic corporate information ecosystem, based on the Universal Document Model proposed by Dagobert Soergel. It focuses on unstructured information objects, especially multimedia, increasingly used in corporate environments, adding semantics and expanding their recovery potential in the composition and reuse of dynamic documents on demand. The proposed model considers stable elements in the organizational environment, such as actors, processes, business metadata and information objects, as well as some basic infrastructures of the corporate information environment. The main objective is to establish a conceptual model that adds semantic intelligence to information assets, leveraging pre-existing infrastructure in organizations, integrating and relating objects to other objects, actors and business processes. The approach methodology considered the state of the art of Information Organization, Representation and Retrieval, Organizational Content Management and Semantic Web technologies, in the scientific literature, as bases for the establishment of an integrative conceptual model. Therefore, the research will be qualitative and exploratory. The predicted steps of the model are: Environment, Data Type and Source Definition, Data Distillation, Metadata Enrichment, and Storage. As a result, in theoretical terms the extended model allows to process heterogeneous and unstructured data according to the established cut-outs and through the processes listed above, allowing value creation in the composition of dynamic information objects, with semantic aggregations to metadata.
  15. Shala, E.: ¬Die Autonomie des Menschen und der Maschine : gegenwärtige Definitionen von Autonomie zwischen philosophischem Hintergrund und technologischer Umsetzbarkeit (2014) 0.02
    0.01668196 = product of:
      0.06672784 = sum of:
        0.06672784 = product of:
          0.20018351 = sum of:
            0.20018351 = weight(_text_:3a in 4388) [ClassicSimilarity], result of:
              0.20018351 = score(doc=4388,freq=2.0), product of:
                0.42742437 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.050415643 = queryNorm
                0.46834838 = fieldWeight in 4388, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4388)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Footnote
    Vgl. unter: https://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=2&cad=rja&uact=8&ved=2ahUKEwizweHljdbcAhVS16QKHXcFD9QQFjABegQICRAB&url=https%3A%2F%2Fwww.researchgate.net%2Fpublication%2F271200105_Die_Autonomie_des_Menschen_und_der_Maschine_-_gegenwartige_Definitionen_von_Autonomie_zwischen_philosophischem_Hintergrund_und_technologischer_Umsetzbarkeit_Redigierte_Version_der_Magisterarbeit_Karls&usg=AOvVaw06orrdJmFF2xbCCp_hL26q.
  16. Piros, A.: Az ETO-jelzetek automatikus interpretálásának és elemzésének kérdései (2018) 0.02
    0.01668196 = product of:
      0.06672784 = sum of:
        0.06672784 = product of:
          0.20018351 = sum of:
            0.20018351 = weight(_text_:3a in 855) [ClassicSimilarity], result of:
              0.20018351 = score(doc=855,freq=2.0), product of:
                0.42742437 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.050415643 = queryNorm
                0.46834838 = fieldWeight in 855, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=855)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Content
    Vgl. auch: New automatic interpreter for complex UDC numbers. Unter: <https%3A%2F%2Fudcc.org%2Ffiles%2FAttilaPiros_EC_36-37_2014-2015.pdf&usg=AOvVaw3kc9CwDDCWP7aArpfjrs5b>
  17. Mattmann, B.: ¬Die Möglichkeiten von RDA bei der Erschliessung historischer Sondermaterialien : Analyse der Beschreibungs- und Erschließungsmöglicjkeiten von Resource Description and Access für die historischen Sondermaterialien Brief und Fotografie (2014) 0.02
    0.016597321 = product of:
      0.066389285 = sum of:
        0.066389285 = weight(_text_:standards in 1651) [ClassicSimilarity], result of:
          0.066389285 = score(doc=1651,freq=2.0), product of:
            0.22470023 = queryWeight, product of:
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.050415643 = queryNorm
            0.29545712 = fieldWeight in 1651, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4569545 = idf(docFreq=1393, maxDocs=44218)
              0.046875 = fieldNorm(doc=1651)
      0.25 = coord(1/4)
    
    Abstract
    Der neue Erschliessungsstandard RDA beschäftigt aktuell unzählige Arbeitsgruppen und Einzelpersonen aus verschiedenen Ländern. Dabei geht es vor allem um die Anpassung einzelner Regeln und Konzepte an lokale Gegebenheiten und spezifische Medientypen. Das erklärte Ziel, den Standard auch für nicht-bibliothekarische Communities attraktiv zu machen, stand jedoch bislang im Hintergrund. Es ist unklar, ob und wie sich RDA beispielsweise in Archiven und Sondersammlungen anwenden lässt. Diese Bachelorarbeit widmet sich dieser Unsicherheit und untersucht die Möglichkeiten der Katalogisierung von ausgewählten archivalischen Ressourcen. Mittels einer Literaturanalyse werden die von Nutzern und Erschliessenden als essenziell erachteten Merkmale von Briefen und Fotografien erhoben und auf die Beschreibungsmöglichkeiten mit RDA hin untersucht. Dabei zeigt sich, dass der Standard grundsätzlich aufgeschlossen gegenüber archivalischen und historischen Ressourcen ist und sich lediglich vereinzelte Anpassungen von Begriffslisten und An- wendungsregeln empfehlen. Entscheidende Fragestellungen, wie die Abbildung von Kontextualität und Hierarchien, müssen jedoch noch geklärt werden, um die Attraktivität des Standards für nicht-bibliothekarische Communities weiter zu verbessern.
  18. Kiren, T.: ¬A clustering based indexing technique of modularized ontologies for information retrieval (2017) 0.02
    0.016541 = product of:
      0.033082 = sum of:
        0.019420752 = weight(_text_:information in 4399) [ClassicSimilarity], result of:
          0.019420752 = score(doc=4399,freq=16.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.21943474 = fieldWeight in 4399, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=4399)
        0.013661247 = product of:
          0.027322493 = sum of:
            0.027322493 = weight(_text_:22 in 4399) [ClassicSimilarity], result of:
              0.027322493 = score(doc=4399,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.15476047 = fieldWeight in 4399, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4399)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Indexing plays a vital role in Information Retrieval. With the availability of huge volume of information, it has become necessary to index the information in such a way to make easier for the end users to find the information they want efficiently and accurately. Keyword-based indexing uses words as indexing terms. It is not capable of capturing the implicit relation among terms or the semantics of the words in the document. To eliminate this limitation, ontology-based indexing came into existence, which allows semantic based indexing to solve complex and indirect user queries. Ontologies are used for document indexing which allows semantic based information retrieval. Existing ontologies or the ones constructed from scratch are used presently for indexing. Constructing ontologies from scratch is a labor-intensive task and requires extensive domain knowledge whereas use of an existing ontology may leave some important concepts in documents un-annotated. Using multiple ontologies can overcome the problem of missing out concepts to a great extent, but it is difficult to manage (changes in ontologies over time by their developers) multiple ontologies and ontology heterogeneity also arises due to ontologies constructed by different ontology developers. One possible solution to managing multiple ontologies and build from scratch is to use modular ontologies for indexing.
    Modular ontologies are built in modular manner by combining modules from multiple relevant ontologies. Ontology heterogeneity also arises during modular ontology construction because multiple ontologies are being dealt with, during this process. Ontologies need to be aligned before using them for modular ontology construction. The existing approaches for ontology alignment compare all the concepts of each ontology to be aligned, hence not optimized in terms of time and search space utilization. A new indexing technique is proposed based on modular ontology. An efficient ontology alignment technique is proposed to solve the heterogeneity problem during the construction of modular ontology. Results are satisfactory as Precision and Recall are improved by (8%) and (10%) respectively. The value of Pearsons Correlation Coefficient for degree of similarity, time, search space requirement, precision and recall are close to 1 which shows that the results are significant. Further research can be carried out for using modular ontology based indexing technique for Multimedia Information Retrieval and Bio-Medical information retrieval.
    Date
    20. 1.2015 18:30:22
  19. Huo, W.: Automatic multi-word term extraction and its application to Web-page summarization (2012) 0.02
    0.015395639 = product of:
      0.030791279 = sum of:
        0.01029941 = weight(_text_:information in 563) [ClassicSimilarity], result of:
          0.01029941 = score(doc=563,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.116372846 = fieldWeight in 563, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=563)
        0.02049187 = product of:
          0.04098374 = sum of:
            0.04098374 = weight(_text_:22 in 563) [ClassicSimilarity], result of:
              0.04098374 = score(doc=563,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.23214069 = fieldWeight in 563, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=563)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    In this thesis we propose three new word association measures for multi-word term extraction. We combine these association measures with LocalMaxs algorithm in our extraction model and compare the results of different multi-word term extraction methods. Our approach is language and domain independent and requires no training data. It can be applied to such tasks as text summarization, information retrieval, and document classification. We further explore the potential of using multi-word terms as an effective representation for general web-page summarization. We extract multi-word terms from human written summaries in a large collection of web-pages, and generate the summaries by aligning document words with these multi-word terms. Our system applies machine translation technology to learn the aligning process from a training set and focuses on selecting high quality multi-word terms from human written summaries to generate suitable results for web-page summarization.
    Date
    10. 1.2013 19:22:47
  20. Lorenz, S.: Konzeption und prototypische Realisierung einer begriffsbasierten Texterschließung (2006) 0.02
    0.015395639 = product of:
      0.030791279 = sum of:
        0.01029941 = weight(_text_:information in 1746) [ClassicSimilarity], result of:
          0.01029941 = score(doc=1746,freq=2.0), product of:
            0.08850355 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.050415643 = queryNorm
            0.116372846 = fieldWeight in 1746, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=1746)
        0.02049187 = product of:
          0.04098374 = sum of:
            0.04098374 = weight(_text_:22 in 1746) [ClassicSimilarity], result of:
              0.04098374 = score(doc=1746,freq=2.0), product of:
                0.17654699 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.050415643 = queryNorm
                0.23214069 = fieldWeight in 1746, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1746)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Im Rahmen dieser Arbeit wird eine Vorgehensweise entwickelt, die die Fixierung auf das Wort und die damit verbundenen Schwächen überwindet. Sie gestattet die Extraktion von Informationen anhand der repräsentierten Begriffe und bildet damit die Basis einer inhaltlichen Texterschließung. Die anschließende prototypische Realisierung dient dazu, die Konzeption zu überprüfen sowie ihre Möglichkeiten und Grenzen abzuschätzen und zu bewerten. Arbeiten zum Information Extraction widmen sich fast ausschließlich dem Englischen, wobei insbesondere im Bereich der Named Entities sehr gute Ergebnisse erzielt werden. Deutlich schlechter sehen die Resultate für weniger regelmäßige Sprachen wie beispielsweise das Deutsche aus. Aus diesem Grund sowie praktischen Erwägungen wie insbesondere der Vertrautheit des Autors damit, soll diese Sprache primär Gegenstand der Untersuchungen sein. Die Lösung von einer engen Termorientierung bei gleichzeitiger Betonung der repräsentierten Begriffe legt nahe, dass nicht nur die verwendeten Worte sekundär werden sondern auch die verwendete Sprache. Um den Rahmen dieser Arbeit nicht zu sprengen wird bei der Untersuchung dieses Punktes das Augenmerk vor allem auf die mit unterschiedlichen Sprachen verbundenen Schwierigkeiten und Besonderheiten gelegt.
    Date
    22. 3.2015 9:17:30

Authors

Languages

  • d 213
  • e 39
  • a 1
  • f 1
  • hu 1
  • pt 1
  • More… Less…

Types