Search (30 results, page 1 of 2)

  • × year_i:[2000 TO 2010}
  • × type_ss:"x"
  1. Stojanovic, N.: Ontology-based Information Retrieval : methods and tools for cooperative query answering (2005) 0.11
    0.11115356 = product of:
      0.2964095 = sum of:
        0.042344213 = product of:
          0.12703264 = sum of:
            0.12703264 = weight(_text_:3a in 701) [ClassicSimilarity], result of:
              0.12703264 = score(doc=701,freq=2.0), product of:
                0.33904418 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.039991006 = queryNorm
                0.3746787 = fieldWeight in 701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
          0.33333334 = coord(1/3)
        0.12703264 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.12703264 = score(doc=701,freq=2.0), product of:
            0.33904418 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.039991006 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
        0.12703264 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.12703264 = score(doc=701,freq=2.0), product of:
            0.33904418 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.039991006 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
      0.375 = coord(3/8)
    
    Content
    Vgl.: http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F1627&ei=tAtYUYrBNoHKtQb3l4GYBw&usg=AFQjCNHeaxKkKU3-u54LWxMNYGXaaDLCGw&sig2=8WykXWQoDKjDSdGtAakH2Q&bvm=bv.44442042,d.Yms.
  2. Munzner, T.: Interactive visualization of large graphs and networks (2000) 0.01
    0.0146638565 = product of:
      0.058655426 = sum of:
        0.023603994 = weight(_text_:computer in 4746) [ClassicSimilarity], result of:
          0.023603994 = score(doc=4746,freq=2.0), product of:
            0.1461475 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.039991006 = queryNorm
            0.16150802 = fieldWeight in 4746, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.03125 = fieldNorm(doc=4746)
        0.03505143 = weight(_text_:network in 4746) [ClassicSimilarity], result of:
          0.03505143 = score(doc=4746,freq=2.0), product of:
            0.17809492 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.039991006 = queryNorm
            0.1968132 = fieldWeight in 4746, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.03125 = fieldNorm(doc=4746)
      0.25 = coord(2/8)
    
    Abstract
    Many real-world domains can be represented as large node-link graphs: backbone Internet routers connect with 70,000 other hosts, mid-sized Web servers handle between 20,000 and 200,000 hyperlinked documents, and dictionaries contain millions of words defined in terms of each other. Computational manipulation of such large graphs is common, but previous tools for graph visualization have been limited to datasets of a few thousand nodes. Visual depictions of graphs and networks are external representations that exploit human visual processing to reduce the cognitive load of many tasks that require understanding of global or local structure. We assert that the two key advantages of computer-based systems for information visualization over traditional paper-based visual exposition are interactivity and scalability. We also argue that designing visualization software by taking the characteristics of a target user's task domain into account leads to systems that are more effective and scale to larger datasets than previous work. This thesis contains a detailed analysis of three specialized systems for the interactive exploration of large graphs, relating the intended tasks to the spatial layout and visual encoding choices. We present two novel algorithms for specialized layout and drawing that use quite different visual metaphors. The H3 system for visualizing the hyperlink structures of web sites scales to datasets of over 100,000 nodes by using a carefully chosen spanning tree as the layout backbone, 3D hyperbolic geometry for a Focus+Context view, and provides a fluid interactive experience through guaranteed frame rate drawing. The Constellation system features a highly specialized 2D layout intended to spatially encode domain-specific information for computational linguists checking the plausibility of a large semantic network created from dictionaries. The Planet Multicast system for displaying the tunnel topology of the Internet's multicast backbone provides a literal 3D geographic layout of arcs on a globe to help MBone maintainers find misconfigured long-distance tunnels. Each of these three systems provides a very different view of the graph structure, and we evaluate their efficacy for the intended task. We generalize these findings in our analysis of the importance of interactivity and specialization for graph visualization systems that are effective and scalable.
  3. Tzitzikas, Y.: Collaborative ontology-based information indexing and retrieval (2002) 0.01
    0.0146638565 = product of:
      0.058655426 = sum of:
        0.023603994 = weight(_text_:computer in 2281) [ClassicSimilarity], result of:
          0.023603994 = score(doc=2281,freq=2.0), product of:
            0.1461475 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.039991006 = queryNorm
            0.16150802 = fieldWeight in 2281, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.03125 = fieldNorm(doc=2281)
        0.03505143 = weight(_text_:network in 2281) [ClassicSimilarity], result of:
          0.03505143 = score(doc=2281,freq=2.0), product of:
            0.17809492 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.039991006 = queryNorm
            0.1968132 = fieldWeight in 2281, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.03125 = fieldNorm(doc=2281)
      0.25 = coord(2/8)
    
    Abstract
    An information system like the Web is a continuously evolving system consisting of multiple heterogeneous information sources, covering a wide domain of discourse, and a huge number of users (human or software) with diverse characteristics and needs, that produce and consume information. The challenge nowadays is to build a scalable information infrastructure enabling the effective, accurate, content based retrieval of information, in a way that adapts to the characteristics and interests of the users. The aim of this work is to propose formally sound methods for building such an information network based on ontologies which are widely used and are easy to grasp by ordinary Web users. The main results of this work are: - A novel scheme for indexing and retrieving objects according to multiple aspects or facets. The proposed scheme is a faceted scheme enriched with a method for specifying the combinations of terms that are valid. We give a model-theoretic interpretation to this model and we provide mechanisms for inferring the valid combinations of terms. This inference service can be exploited for preventing errors during the indexing process, which is very important especially in the case where the indexing is done collaboratively by many users, and for deriving "complete" navigation trees suitable for browsing through the Web. The proposed scheme has several advantages over the hierarchical classification schemes currently employed by Web catalogs, namely, conceptual clarity (it is easier to understand), compactness (it takes less space), and scalability (the update operations can be formulated more easily and be performed more effciently). - A exible and effecient model for building mediators over ontology based information sources. The proposed mediators support several modes of query translation and evaluation which can accommodate various application needs and levels of answer quality. The proposed model can be used for providing users with customized views of Web catalogs. It can also complement the techniques for building mediators over relational sources so as to support approximate translation of partially ordered domain values.
    Imprint
    Heraklion : University of Crete / Department of Computer Science
  4. Lehrke, C.: Architektur von Suchmaschinen : Googles Architektur, insb. Crawler und Indizierer (2005) 0.01
    0.010762642 = product of:
      0.04305057 = sum of:
        0.029504994 = weight(_text_:computer in 867) [ClassicSimilarity], result of:
          0.029504994 = score(doc=867,freq=2.0), product of:
            0.1461475 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.039991006 = queryNorm
            0.20188503 = fieldWeight in 867, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0390625 = fieldNorm(doc=867)
        0.013545574 = product of:
          0.027091147 = sum of:
            0.027091147 = weight(_text_:22 in 867) [ClassicSimilarity], result of:
              0.027091147 = score(doc=867,freq=2.0), product of:
                0.1400417 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.039991006 = queryNorm
                0.19345059 = fieldWeight in 867, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=867)
          0.5 = coord(1/2)
      0.25 = coord(2/8)
    
    Abstract
    Das Internet mit seinen ständig neuen Usern und seinem extremen Wachstum bringt viele neue Herausforderungen mit sich. Aufgrund dieses Wachstums bedienen sich die meisten Leute der Hilfe von Suchmaschinen um Inhalte innerhalb des Internet zu finden. Suchmaschinen nutzen für die Beantwortung der User-Anfragen Information Retrieval Techniken. Problematisch ist nur, dass traditionelle Information Retrieval (IR) Systeme für eine relativ kleine und zusammenhängende Sammlung von Dokumenten entwickelt wurden. Das Internet hingegen unterliegt einem ständigen Wachstum, schnellen Änderungsraten und es ist über geographisch verteilte Computer verteilt. Aufgrund dieser Tatsachen müssen die alten Techniken erweitert oder sogar neue IRTechniken entwickelt werden. Eine Suchmaschine die diesen Herausforderungen vergleichsweise erfolgreich entgegnet ist Google. Ziel dieser Arbeit ist es aufzuzeigen, wie Suchmaschinen funktionieren. Der Fokus liegt dabei auf der Suchmaschine Google. Kapitel 2 wird sich zuerst mit dem Aufbau von Suchmaschinen im Allgemeinen beschäftigen, wodurch ein grundlegendes Verständnis für die einzelnen Komponenten geschaffen werden soll. Im zweiten Teil des Kapitels wird darauf aufbauend ein Überblick über die Architektur von Google gegeben. Kapitel 3 und 4 dienen dazu, näher auf die beiden Komponenten Crawler und Indexer einzugehen, bei denen es sich um zentrale Elemente im Rahmen von Suchmaschinen handelt.
    Pages
    22 S
  5. Quosig, D.: Umsetzung des Lehrbuches "Wirtschaftsinformation" in ein Online-Tutorial (2004) 0.01
    0.010326747 = product of:
      0.082613975 = sum of:
        0.082613975 = weight(_text_:computer in 4527) [ClassicSimilarity], result of:
          0.082613975 = score(doc=4527,freq=2.0), product of:
            0.1461475 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.039991006 = queryNorm
            0.56527805 = fieldWeight in 4527, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.109375 = fieldNorm(doc=4527)
      0.125 = coord(1/8)
    
    Theme
    Computer Based Training
  6. Praetsch, I.: ¬Die Bereitstellung von digitalen Lehrmaterialien im Content Management System des Fachbereiches Informationswissenschaften an der Fachhochschule Potsdam exemplarisch an der Lehrveranstaltung 'Internet- und Webtechnologie' (2004) 0.01
    0.008851498 = product of:
      0.07081199 = sum of:
        0.07081199 = weight(_text_:computer in 4634) [ClassicSimilarity], result of:
          0.07081199 = score(doc=4634,freq=2.0), product of:
            0.1461475 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.039991006 = queryNorm
            0.48452407 = fieldWeight in 4634, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.09375 = fieldNorm(doc=4634)
      0.125 = coord(1/8)
    
    Theme
    Computer Based Training
  7. Kumpe, D.: Methoden zur automatischen Indexierung von Dokumenten (2006) 0.01
    0.0076675005 = product of:
      0.061340004 = sum of:
        0.061340004 = weight(_text_:network in 782) [ClassicSimilarity], result of:
          0.061340004 = score(doc=782,freq=2.0), product of:
            0.17809492 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.039991006 = queryNorm
            0.3444231 = fieldWeight in 782, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.0546875 = fieldNorm(doc=782)
      0.125 = coord(1/8)
    
    Abstract
    Diese Diplomarbeit handelt von der Indexierung von unstrukturierten und natürlichsprachigen Dokumenten. Die zunehmende Informationsflut und die Zahl an veröffentlichten wissenschaftlichen Berichten und Büchern machen eine maschinelle inhaltliche Erschließung notwendig. Um die Anforderungen hierfür besser zu verstehen, werden Probleme der natürlichsprachigen schriftlichen Kommunikation untersucht. Die manuellen Techniken der Indexierung und die Dokumentationssprachen werden vorgestellt. Die Indexierung wird thematisch in den Bereich der inhaltlichen Erschließung und des Information Retrieval eingeordnet. Weiterhin werden Vor- und Nachteile von ausgesuchten Algorithmen untersucht und Softwareprodukte im Bereich des Information Retrieval auf ihre Arbeitsweise hin evaluiert. Anhand von Beispiel-Dokumenten werden die Ergebnisse einzelner Verfahren vorgestellt. Mithilfe des Projekts European Migration Network werden Probleme und grundlegende Anforderungen an die Durchführung einer inhaltlichen Erschließung identifiziert und Lösungsmöglichkeiten vorgeschlagen.
  8. Buß, M.: Unternehmenssprache in internationalen Unternehmen : Probleme des Informationstransfers in der internen Kommunikation (2005) 0.01
    0.0070661027 = product of:
      0.05652882 = sum of:
        0.05652882 = sum of:
          0.029437674 = weight(_text_:resources in 1482) [ClassicSimilarity], result of:
            0.029437674 = score(doc=1482,freq=2.0), product of:
              0.14598069 = queryWeight, product of:
                3.650338 = idf(docFreq=3122, maxDocs=44218)
                0.039991006 = queryNorm
              0.20165458 = fieldWeight in 1482, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.650338 = idf(docFreq=3122, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1482)
          0.027091147 = weight(_text_:22 in 1482) [ClassicSimilarity], result of:
            0.027091147 = score(doc=1482,freq=2.0), product of:
              0.1400417 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.039991006 = queryNorm
              0.19345059 = fieldWeight in 1482, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1482)
      0.125 = coord(1/8)
    
    Date
    22. 5.2005 18:25:26
    Theme
    Information Resources Management
  9. Makewita, S.M.: Investigating the generic information-seeking function of organisational decision-makers : perspectives on improving organisational information systems (2002) 0.01
    0.0070661027 = product of:
      0.05652882 = sum of:
        0.05652882 = sum of:
          0.029437674 = weight(_text_:resources in 642) [ClassicSimilarity], result of:
            0.029437674 = score(doc=642,freq=2.0), product of:
              0.14598069 = queryWeight, product of:
                3.650338 = idf(docFreq=3122, maxDocs=44218)
                0.039991006 = queryNorm
              0.20165458 = fieldWeight in 642, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.650338 = idf(docFreq=3122, maxDocs=44218)
                0.0390625 = fieldNorm(doc=642)
          0.027091147 = weight(_text_:22 in 642) [ClassicSimilarity], result of:
            0.027091147 = score(doc=642,freq=2.0), product of:
              0.1400417 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.039991006 = queryNorm
              0.19345059 = fieldWeight in 642, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=642)
      0.125 = coord(1/8)
    
    Abstract
    The past decade has seen the emergence of a new paradigm in the corporate world where organisations emphasised connectivity as a means of exposing decision-makers to wider resources of information within and outside the organisation. Many organisations followed the initiatives of enhancing infrastructures, manipulating cultural shifts and emphasising managerial commitment for creating pools and networks of knowledge. However, the concept of connectivity is not merely presenting people with the data, but more importantly, to create environments where people can seek information efficiently. This paradigm has therefore caused a shift in the function of information systems in organisations. They have to be now assessed in relation to how they underpin people's information-seeking activities within the context of their organisational environment. This research project used interpretative research methods to investigate the nature of people's information-seeking activities at two culturally contrasting organisations. Outcomes of this research project provide insights into phenomena associated with people's information-seeking function, and show how they depend on the organisational context that is defined partly by information systems. It suggests that information-seeking is not just searching for data. The inefficiencies inherent in both people and their environments can bring opaqueness into people's data, which they need to avoid or eliminate as part of seeking information. This seems to have made information-seeking a two-tier process consisting of a primary process of searching and interpreting data and auxiliary process of avoiding and eliminating opaqueness in data. Based on this view, this research suggests that organisational information systems operate naturally as implicit dual-mechanisms to underpin the above two-tier process, and that improvements to information systems should concern maintaining the balance in these dual-mechanisms.
    Date
    22. 7.2022 12:16:58
  10. Kuhlwein, K.: E-Learning - ein Aufgabenfeld für Bibliothekare? : Eine Untersuchung ausgewählter Online-Schulungskonzepte an deutschen Bibliotheken (2004) 0.01
    0.0059009986 = product of:
      0.04720799 = sum of:
        0.04720799 = weight(_text_:computer in 3580) [ClassicSimilarity], result of:
          0.04720799 = score(doc=3580,freq=2.0), product of:
            0.1461475 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.039991006 = queryNorm
            0.32301605 = fieldWeight in 3580, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0625 = fieldNorm(doc=3580)
      0.125 = coord(1/8)
    
    Theme
    Computer Based Training
  11. Lipokatic, R.: Vergleichende Usability-Evaluation der E-Learning Plattformen LUVIT und WebCT (2004) 0.01
    0.0059009986 = product of:
      0.04720799 = sum of:
        0.04720799 = weight(_text_:computer in 3702) [ClassicSimilarity], result of:
          0.04720799 = score(doc=3702,freq=2.0), product of:
            0.1461475 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.039991006 = queryNorm
            0.32301605 = fieldWeight in 3702, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0625 = fieldNorm(doc=3702)
      0.125 = coord(1/8)
    
    Theme
    Computer Based Training
  12. Haslhofer, B.: ¬A Web-based mapping technique for establishing metadata interoperability (2008) 0.01
    0.005542328 = product of:
      0.044338625 = sum of:
        0.044338625 = weight(_text_:property in 3173) [ClassicSimilarity], result of:
          0.044338625 = score(doc=3173,freq=2.0), product of:
            0.25336683 = queryWeight, product of:
              6.335595 = idf(docFreq=212, maxDocs=44218)
              0.039991006 = queryNorm
            0.17499775 = fieldWeight in 3173, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.335595 = idf(docFreq=212, maxDocs=44218)
              0.01953125 = fieldNorm(doc=3173)
      0.125 = coord(1/8)
    
    Abstract
    The integration of metadata from distinct, heterogeneous data sources requires metadata interoperability, which is a qualitative property of metadata information objects that is not given by default. The technique of metadata mapping allows domain experts to establish metadata interoperability in a certain integration scenario. Mapping solutions, as a technical manifestation of this technique, are already available for the intensively studied domain of database system interoperability, but they rarely exist for the Web. If we consider the amount of steadily increasing structured metadata and corresponding metadata schemes on theWeb, we can observe a clear need for a mapping solution that can operate in aWeb-based environment. To achieve that, we first need to build its technical core, which is a mapping model that provides the language primitives to define mapping relationships. Existing SemanticWeb languages such as RDFS and OWL define some basic mapping elements (e.g., owl:equivalentProperty, owl:sameAs), but do not address the full spectrum of semantic and structural heterogeneities that can occur among distinct, incompatible metadata information objects. Furthermore, it is still unclear how to process defined mapping relationships during run-time in order to deliver metadata to the client in a uniform way. As the main contribution of this thesis, we present an abstract mapping model, which reflects the mapping problem on a generic level and provides the means for reconciling incompatible metadata. Instance transformation functions and URIs take a central role in that model. The former cover a broad spectrum of possible structural and semantic heterogeneities, while the latter bind the complete mapping model to the architecture of the Word Wide Web. On the concrete, language-specific level we present a binding of the abstract mapping model for the RDF Vocabulary Description Language (RDFS), which allows us to create mapping specifications among incompatible metadata schemes expressed in RDFS. The mapping model is embedded in a cyclic process that categorises the requirements a mapping solution should fulfil into four subsequent phases: mapping discovery, mapping representation, mapping execution, and mapping maintenance. In this thesis, we mainly focus on mapping representation and on the transformation of mapping specifications into executable SPARQL queries. For mapping discovery support, the model provides an interface for plugging-in schema and ontology matching algorithms. For mapping maintenance we introduce the concept of a simple, but effective mapping registry. Based on the mapping model, we propose aWeb-based mediator wrapper-architecture that allows domain experts to set up mediation endpoints that provide a uniform SPARQL query interface to a set of distributed metadata sources. The involved data sources are encapsulated by wrapper components that expose the contained metadata and the schema definitions on the Web and provide a SPARQL query interface to these metadata. In this thesis, we present the OAI2LOD Server, a wrapper component for integrating metadata that are accessible via the Open Archives Initiative Protocol for Metadata Harvesting (OAI-PMH). In a case study, we demonstrate how mappings can be created in aWeb environment and how our mediator wrapper architecture can easily be configured in order to integrate metadata from various heterogeneous data sources without the need to install any mapping solution or metadata integration solution in a local system environment.
  13. Weidler, D.: Usability-Inspektionen und Usability-Tests : Komplementäre Methoden für die Evaluation eines E-Learning-Moduls (2004) 0.01
    0.0051633734 = product of:
      0.041306987 = sum of:
        0.041306987 = weight(_text_:computer in 3706) [ClassicSimilarity], result of:
          0.041306987 = score(doc=3706,freq=2.0), product of:
            0.1461475 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.039991006 = queryNorm
            0.28263903 = fieldWeight in 3706, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3706)
      0.125 = coord(1/8)
    
    Theme
    Computer Based Training
  14. Sperling, R.: Anlage von Literaturreferenzen für Onlineressourcen auf einer virtuellen Lernplattform (2004) 0.00
    0.0047409507 = product of:
      0.037927605 = sum of:
        0.037927605 = product of:
          0.07585521 = sum of:
            0.07585521 = weight(_text_:22 in 4635) [ClassicSimilarity], result of:
              0.07585521 = score(doc=4635,freq=2.0), product of:
                0.1400417 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.039991006 = queryNorm
                0.5416616 = fieldWeight in 4635, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4635)
          0.5 = coord(1/2)
      0.125 = coord(1/8)
    
    Date
    26.11.2005 18:39:22
  15. Milanesi, C.: Möglichkeiten der Kooperation im Rahmen von Subject Gateways : das Euler-Projekt im Vergleich mit weiteren europäischen Projekten (2001) 0.00
    0.004063672 = product of:
      0.032509375 = sum of:
        0.032509375 = product of:
          0.06501875 = sum of:
            0.06501875 = weight(_text_:22 in 4865) [ClassicSimilarity], result of:
              0.06501875 = score(doc=4865,freq=2.0), product of:
                0.1400417 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.039991006 = queryNorm
                0.46428138 = fieldWeight in 4865, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4865)
          0.5 = coord(1/2)
      0.125 = coord(1/8)
    
    Date
    22. 6.2002 19:41:59
  16. Haller, S.H.M.: Mappingverfahren zur Wissensorganisation (2002) 0.00
    0.0033863934 = product of:
      0.027091147 = sum of:
        0.027091147 = product of:
          0.054182295 = sum of:
            0.054182295 = weight(_text_:22 in 3406) [ClassicSimilarity], result of:
              0.054182295 = score(doc=3406,freq=2.0), product of:
                0.1400417 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.039991006 = queryNorm
                0.38690117 = fieldWeight in 3406, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3406)
          0.5 = coord(1/2)
      0.125 = coord(1/8)
    
    Date
    30. 5.2010 16:22:35
  17. Bachfeld, S.: Möglichkeiten und Grenzen linguistischer Verfahren der automatischen Indexierung : Entwurf einer Simulation für den Einsatz im Grundstudium (2003) 0.00
    0.0029504993 = product of:
      0.023603994 = sum of:
        0.023603994 = weight(_text_:computer in 2827) [ClassicSimilarity], result of:
          0.023603994 = score(doc=2827,freq=2.0), product of:
            0.1461475 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.039991006 = queryNorm
            0.16150802 = fieldWeight in 2827, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.03125 = fieldNorm(doc=2827)
      0.125 = coord(1/8)
    
    Theme
    Computer Based Training
  18. Bierbach, P.: Wissensrepräsentation - Gegenstände und Begriffe : Bedingungen des Antinomieproblems bei Frege und Chancen des Begriffssystems bei Lambert (2001) 0.00
    0.0029504993 = product of:
      0.023603994 = sum of:
        0.023603994 = weight(_text_:computer in 4498) [ClassicSimilarity], result of:
          0.023603994 = score(doc=4498,freq=2.0), product of:
            0.1461475 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.039991006 = queryNorm
            0.16150802 = fieldWeight in 4498, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.03125 = fieldNorm(doc=4498)
      0.125 = coord(1/8)
    
    Abstract
    Die auf Basis vernetzter Computer realisierbare Möglichkeit einer universalen Enzyklopädie führt aufgrund der dabei technisch notwendigen Reduktion auf nur eine Sorte Repräsentanten zu Systemen, bei denen entweder nur Gegenstände repräsentiert werden, die auch Begriffe vertreten, oder nur Begriffe, die auch Gegenstände vertreten. In der Dissertation werden als Beispiele solcher Repräsentationssysteme die logischen Systeme von Gottlob Frege und Johann Heinrich Lambert untersucht. Freges System, basierend auf der Annahme der Objektivität von Bedeutungen, war durch die Nachweisbarkeit einer Antinomie gescheitert, weshalb von Philosophen im 20. Jahrhundert die Existenz einer objektiven Bedeutung von Ausdrücken und die Übersetzbarkeit der Gedanken aus den natürlichen Sprachen in eine formale Sprache in Frage gestellt wurde. In der Dissertation wird nachgewiesen, daß diese Konsequenz voreilig war und daß die Antinomie auch bei Annahme der Objektivität von Wissen erst durch zwei Zusatzforderungen in Freges Logik ausgelöst wird: die eineindeutige Zuordnung eines Gegenstands zu jedem Begriff sowie die scharfen Begrenzung der Begriffe, die zur Abgeschlossenheit des Systems zwingt. Als Alternative wird das Begriffssystem Lamberts diskutiert, bei dem jeder Gegenstand durch einen Begriff und gleichwertig durch Gesamtheiten von Begriffen vertreten wird und Begriffe durch Gesamtheiten von Begriffen ersetzbar sind. Beide die Antinomie auslösenden Bedingungen sind hier nicht vorhanden, zugleich ist die fortschreitende Entwicklung von Wissen repräsentierbar. Durch die mengentheoretische Rekonstruktion des Begriffssystems Lamberts in der Dissertation wird dessen praktische Nutzbarkeit gezeigt. Resultat der Dissertation ist der Nachweis, daß es Repräsentationssysteme gibt, die nicht auf die für die Prüfung der Verbindlichkeit der Einträge in die Enzyklopädie notwendige Annahme der Verobjektivierbarkeit von Wissen verzichten müssen, weil ihnen nicht jene die Antinomie auslösenden Voraussetzungen zugrunde liegen.
  19. Styltsvig, H.B.: Ontology-based information retrieval (2006) 0.00
    0.0029504993 = product of:
      0.023603994 = sum of:
        0.023603994 = weight(_text_:computer in 1154) [ClassicSimilarity], result of:
          0.023603994 = score(doc=1154,freq=2.0), product of:
            0.1461475 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.039991006 = queryNorm
            0.16150802 = fieldWeight in 1154, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.03125 = fieldNorm(doc=1154)
      0.125 = coord(1/8)
    
    Imprint
    Roskilde : Roskilde University, Computer Science Section
  20. Hoffmann, R.: Mailinglisten für den bibliothekarischen Informationsdienst am Beispiel von RABE (2000) 0.00
    0.0028734503 = product of:
      0.022987602 = sum of:
        0.022987602 = product of:
          0.045975205 = sum of:
            0.045975205 = weight(_text_:22 in 4441) [ClassicSimilarity], result of:
              0.045975205 = score(doc=4441,freq=4.0), product of:
                0.1400417 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.039991006 = queryNorm
                0.32829654 = fieldWeight in 4441, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4441)
          0.5 = coord(1/2)
      0.125 = coord(1/8)
    
    Date
    22. 2.2000 10:25:05
    Series
    Kölner Arbeitspapiere zur Bibliotheks- und Informationswissenschaft; Bd.22