Search (54 results, page 1 of 3)

  • × theme_ss:"Verteilte bibliographische Datenbanken"
  1. Dempsey, L.; Russell, R.; Kirriemur, J.W.: Towards distributed library systems : Z39.50 in a European context (1996) 0.07
    0.06633483 = product of:
      0.099502236 = sum of:
        0.03559984 = weight(_text_:retrieval in 127) [ClassicSimilarity], result of:
          0.03559984 = score(doc=127,freq=2.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.26736724 = fieldWeight in 127, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=127)
        0.06390239 = product of:
          0.09585358 = sum of:
            0.04814333 = weight(_text_:29 in 127) [ClassicSimilarity], result of:
              0.04814333 = score(doc=127,freq=2.0), product of:
                0.15484026 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04401763 = queryNorm
                0.31092256 = fieldWeight in 127, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=127)
            0.04771025 = weight(_text_:22 in 127) [ClassicSimilarity], result of:
              0.04771025 = score(doc=127,freq=2.0), product of:
                0.15414225 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04401763 = queryNorm
                0.30952093 = fieldWeight in 127, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=127)
          0.6666667 = coord(2/3)
      0.6666667 = coord(2/3)
    
    Abstract
    Z39.50 is an information retrieval protocol. It has generated much interest but is so far little deployed in UK systems and services. Gives a functional overview of the protocol itself and the standards background, describes some European initiatives which make use of it, and outlines various issues to do with its future use and acceptance. Z39.50 is a crucial building block of future distributed information systems but it needs to be considered alongside other protocols and services to provide useful applications
    Date
    3. 3.1999 17:29:59
    Source
    Program. 30(1996) no.1, S.1-22
  2. Friedrich, M.; Schimkat, R.-D.; Küchlin, W.: Information retrieval in distributed environments based on context-aware, proactive documents (2002) 0.04
    0.038729575 = product of:
      0.05809436 = sum of:
        0.044052552 = weight(_text_:retrieval in 3608) [ClassicSimilarity], result of:
          0.044052552 = score(doc=3608,freq=4.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.33085006 = fieldWeight in 3608, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3608)
        0.014041806 = product of:
          0.042125415 = sum of:
            0.042125415 = weight(_text_:29 in 3608) [ClassicSimilarity], result of:
              0.042125415 = score(doc=3608,freq=2.0), product of:
                0.15484026 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04401763 = queryNorm
                0.27205724 = fieldWeight in 3608, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3608)
          0.33333334 = coord(1/3)
      0.6666667 = coord(2/3)
    
    Abstract
    In this position paper we propose a document-centric middleware component called Living Documents to support context-aware information retrieval in distributed communities. A Living Document acts as a micro server for a document which contains computational services, a semi-structured knowledge repository to uniformly store and access context-related information, and finally the document's digital content. Our initial prototype of Living Documents is based an the concept of mobile agents and implemented in Java and XML.
    Source
    Gaining insight from research information (CRIS2002): Proceedings of the 6th International Conference an Current Research Information Systems, University of Kassel, August 29 - 31, 2002. Eds: W. Adamczak u. A. Nase
  3. Lopatenko, A.; Asserson, A.; Jeffery, K.G.: CERIF - Information retrieval of research information in a distributed heterogeneous environment (2002) 0.03
    0.033196777 = product of:
      0.049795166 = sum of:
        0.037759334 = weight(_text_:retrieval in 3597) [ClassicSimilarity], result of:
          0.037759334 = score(doc=3597,freq=4.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.2835858 = fieldWeight in 3597, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=3597)
        0.012035834 = product of:
          0.0361075 = sum of:
            0.0361075 = weight(_text_:29 in 3597) [ClassicSimilarity], result of:
              0.0361075 = score(doc=3597,freq=2.0), product of:
                0.15484026 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04401763 = queryNorm
                0.23319192 = fieldWeight in 3597, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3597)
          0.33333334 = coord(1/3)
      0.6666667 = coord(2/3)
    
    Abstract
    User demands to have access to complete and actual information about research may require integration of data from different CRISs. CRISs are rarely homogenous systems and problems of CRISs integration must be addressed from technological point of view. Implementation of CRIS providing access to heterogeneous data distributed among a number of CRISs is described. A few technologies - distributed databases, web services, semantic web are used for distributed CRIS to address different user requirements. Distributed databases serve to implement very efficient integration of homogenous systems, web services - to provide open access to research information, semantic web - to solve problems of integration semantically and structurally heterogeneous data sources and provide intelligent data retrieval interfaces. The problems of data completeness in distributed systems are addressed and CRIS-adequate solution for data completeness is suggested.
    Source
    Gaining insight from research information (CRIS2002): Proceedings of the 6th International Conference an Current Research Information Systems, University of Kassel, August 29 - 31, 2002. Eds: W. Adamczak u. A. Nase
  4. Sarinder, K.K.S.; Lim, L.H.S.; Merican, A.F.; Dimyati, K.: Biodiversity information retrieval across networked data sets (2010) 0.03
    0.030433651 = product of:
      0.045650475 = sum of:
        0.031466108 = weight(_text_:retrieval in 3951) [ClassicSimilarity], result of:
          0.031466108 = score(doc=3951,freq=4.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.23632148 = fieldWeight in 3951, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3951)
        0.014184365 = product of:
          0.042553093 = sum of:
            0.042553093 = weight(_text_:29 in 3951) [ClassicSimilarity], result of:
              0.042553093 = score(doc=3951,freq=4.0), product of:
                0.15484026 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04401763 = queryNorm
                0.2748193 = fieldWeight in 3951, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3951)
          0.33333334 = coord(1/3)
      0.6666667 = coord(2/3)
    
    Abstract
    Purpose - Biodiversity resources are inevitably digital and stored in a wide variety of formats by researchers or stakeholders. In Malaysia, although digitizing biodiversity data has long been stressed, the interoperability of the biodiversity data is still an issue that requires attention. This is because, when data are shared, the question of copyright occurs, creating a setback among researchers wanting to promote or share data through online presentations. To solve this, the aim is to present an approach to integrate data through wrapping of datasets stored in relational databases located on networked platforms. Design/methodology/approach - The approach uses tools such as XML, PHP, ASP and HTML to integrate distributed databases in heterogeneous formats. Five current database integration systems were reviewed and all of them have common attributes such as query-oriented, using a mediator-based approach and integrating a structured data model. These common attributes were also adopted in the proposed solution. Distributed Generic Information Retrieval (DiGIR) was used as a model in designing the proposed solution. Findings - A new database integration system was developed, which is user-friendly and simple with common attributes found in current integration systems.
    Date
    29. 8.2010 12:29:10
  5. Stark, T.: ¬The Net and Z39.50 : toward a virtual union catalog (1997) 0.03
    0.030127779 = product of:
      0.045191668 = sum of:
        0.03114986 = weight(_text_:retrieval in 3194) [ClassicSimilarity], result of:
          0.03114986 = score(doc=3194,freq=2.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.23394634 = fieldWeight in 3194, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3194)
        0.014041806 = product of:
          0.042125415 = sum of:
            0.042125415 = weight(_text_:29 in 3194) [ClassicSimilarity], result of:
              0.042125415 = score(doc=3194,freq=2.0), product of:
                0.15484026 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04401763 = queryNorm
                0.27205724 = fieldWeight in 3194, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3194)
          0.33333334 = coord(1/3)
      0.6666667 = coord(2/3)
    
    Abstract
    The State Library of Iowa, USA, received a Higher Education Act title II grant from the US Dept. of Education in 1994 to create a demonstration project of new library information technologies. Describes 2 interlinked components of the project: Web-based union catalogue development and statewide deployment of the ANSI/NISO Z39.50 standard for database search and retrieval. Z39.50 was chosen because of its ability to searching multiple remote databases in a single session and its common interface across a variety of implementations. Use of a distributed Z39.50 search makes the need for maintaining large union catalogues unnecessary
    Source
    Computers in libraries. 17(1997) no.10, S.27-29
  6. Johnson, E.H.: Objects for distributed heterogeneous information retrieval (2000) 0.03
    0.027603831 = product of:
      0.041405745 = sum of:
        0.031466108 = weight(_text_:retrieval in 6959) [ClassicSimilarity], result of:
          0.031466108 = score(doc=6959,freq=4.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.23632148 = fieldWeight in 6959, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6959)
        0.009939636 = product of:
          0.029818907 = sum of:
            0.029818907 = weight(_text_:22 in 6959) [ClassicSimilarity], result of:
              0.029818907 = score(doc=6959,freq=2.0), product of:
                0.15414225 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04401763 = queryNorm
                0.19345059 = fieldWeight in 6959, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=6959)
          0.33333334 = coord(1/3)
      0.6666667 = coord(2/3)
    
    Abstract
    The success of the World Wide Web Shows that we can access, search, and retrieve information from globally distributed databases. lf a database, such as a library catalog, has some sort of Web-based front end, we can type its URL into a Web browser and use its HTML-based forms to search for items in that database. Depending an how well the query conforms to the database content, how the search engine interprets the query, and how the server formats the results into HTML, we might actually find something usable. While the first two issues depend an ourselves and the server, an the Web the latter falls to the mercy of HTML, which we all know as a great destroyer of information because it codes for display but not for content description. When looking at an HTML-formatted display, we must depend an our own interpretation to recognize such entities as author names, titles, and subject identifiers. The Web browser can do nothing but display the information. lf we want some other view of the result, such as sorting the records by date (provided it offers such an option to begin with), the server must do it. This makes poor use of the computing power we have at the desktop (or even laptop), which, unless it involves retrieving more records, could easily do the result Set manipulation that we currently send back to the server. Despite having personal computers wich immense computational power, as far as information retrieval goes, we still essentially use them as dumb terminals.
    Date
    22. 9.1997 19:16:05
  7. Xu, J.; Croft, W.B.: Topic-based language models for distributed retrieval (2000) 0.03
    0.026699878 = product of:
      0.080099635 = sum of:
        0.080099635 = weight(_text_:retrieval in 38) [ClassicSimilarity], result of:
          0.080099635 = score(doc=38,freq=18.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.60157627 = fieldWeight in 38, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=38)
      0.33333334 = coord(1/3)
    
    Abstract
    Effective retrieval in a distributed environment is an important but difficult problem. Lack of effectiveness appears to have two major causes. First, existing collection selection algorithms do not work well on heterogeneous collections. Second, relevant documents are scattered over many collections and searching a few collections misses many relevant documents. We propose a topic-oriented approach to distributed retrieval. With this approach, we structure the document set of a distributed retrieval environment around a set of topics. Retrieval for a query involves first selecting the right topics for the query and then dispatching the search process to collections that contain such topics. The content of a topic is characterized by a language model. In environments where the labeling of documents by topics is unavailable, document clustering is employed for topic identification. Based on these ideas, three methods are proposed to suit different environments. We show that all three methods improve effectiveness of distributed retrieval
    Series
    The Kluwer international series on information retrieval; 7
    Source
    Advances in information retrieval: Recent research from the Center for Intelligent Information Retrieval. Ed.: W.B. Croft
  8. Tappenbeck, I.; Wessel, C.: CARMEN : Content Analysis, Retrieval and Metadata: Effective Net-working. Ein Halbzeitbericht (2001) 0.03
    0.025902838 = product of:
      0.038854256 = sum of:
        0.030830367 = weight(_text_:retrieval in 5900) [ClassicSimilarity], result of:
          0.030830367 = score(doc=5900,freq=6.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.23154683 = fieldWeight in 5900, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=5900)
        0.008023889 = product of:
          0.024071665 = sum of:
            0.024071665 = weight(_text_:29 in 5900) [ClassicSimilarity], result of:
              0.024071665 = score(doc=5900,freq=2.0), product of:
                0.15484026 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04401763 = queryNorm
                0.15546128 = fieldWeight in 5900, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5900)
          0.33333334 = coord(1/3)
      0.6666667 = coord(2/3)
    
    Abstract
    Das Projekt CARMEN startete als Sonderfördermaßnahme im Rahmen von Global lnfo im Oktober 1999 mit einer geplanten Laufzeit von 29 Monaten. Der Schwerpunkt des Projekts liegt in der Weiterentwicklung von Konzepten und Verfahren der Dokumenterschließung, die den Zugriff auf heterogene, dezentral verteilte Informationsbestände und deren Verwaltung nach gemeinsamen Prinzipien ermöglichen sollen. Dabei geht CARMEN gezielt einen anderen Weg als die meisten bisherigen Ansätze in diesem Bereich, die versuchen, Homogenität und Konsistenz in einer dezentralen Informationslandschaft technikorientiert herzustellen, indem Verfahren entwickelt werden, durch die physikalisch auf verschiedene Dokumentenräume gleichzeitig zugegriffen werden kann. Eine rein technische Parallelisierung von Zugriffsmöglichkeiten reicht jedoch nicht aus, denn das Hauptproblem der inhaltlichen, strukturellen und konzeptionellen Differenz der einzelnen Datenbestände wird damit nicht gelöst. Um diese Differenzen zu kompensieren, werden Problemlösungen und Weiterentwicklungen innerhalb des Projekts CARMEN in drei Bereichen erarbeitet: (1) Metadaten (Dokumentbeschreibung, Retrieval, Verwaltung, Archivierung) (2) Methoden des Umgangs mit der verbleibenden Heterogenität der Datenbestände (3) Retrieval für strukturierte Dokumente mit Metadaten und heterogenen Datentypen. Diese drei Aufgabenbereiche hängen eng zusammen. Durch die Entwicklungen im Bereich der Metadaten soll einerseits die verlorengegangene Konsistenz partiell wiederhergestellt und auf eine den neuen Medien gerechte Basis gestellt werden. Andererseits sollen durch Verfahren zur Heterogenitätsbehandlung Dokumente mit unterschiedlicher Datenrelevanz und Inhaltserschließung aufeinander bezogen und retrievalseitig durch ein Rechercheverfahren erganzt werden, das den unterschiedlichen Datentypen gerecht wird Innerhalb des Gesamtprojekts CARMEN werden diese Aspekte arbeitsteilig behandelt. Acht Arbeitspakete (APs) befassen sich in Abstimmung miteinander mit je verschiedenen Schwerpunkten. Um die Koordination der Arbeiten der verschiedenen APs untereinander zu unterstützen, trafen sich die ca. 40 Projektbearbeiter am 1. und 2. Februar 2001 zum "CARMEN middle OfTheRoad Workshop" in Bonn. Anlässlich dieses Workshops wurden die inhaltlichen und technischen Ergebnisse, die in der ersten Hälfte der Projektlaufzeit von den einzelnen APs erzielt worden sind, in insgesamt 17 Präsentationen vorgestellt
  9. Tappenbeck, I.; Wessel, C.: CARMEN : Content Analysis, Retrieval and Metadata: Effective Net-working. Bericht über den middleOfTheRoad Workshop (2001) 0.03
    0.025902838 = product of:
      0.038854256 = sum of:
        0.030830367 = weight(_text_:retrieval in 5901) [ClassicSimilarity], result of:
          0.030830367 = score(doc=5901,freq=6.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.23154683 = fieldWeight in 5901, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=5901)
        0.008023889 = product of:
          0.024071665 = sum of:
            0.024071665 = weight(_text_:29 in 5901) [ClassicSimilarity], result of:
              0.024071665 = score(doc=5901,freq=2.0), product of:
                0.15484026 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04401763 = queryNorm
                0.15546128 = fieldWeight in 5901, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5901)
          0.33333334 = coord(1/3)
      0.6666667 = coord(2/3)
    
    Abstract
    Das Projekt CARMEN startete als Sonderfördermaßnahme im Rahmen von Global lnfo im Oktober 1999 mit einer geplanten Laufzeit von 29 Monaten. Der Schwerpunkt des Projekts liegt in der Weiterentwicklung von Konzepten und Verfahren der Dokumenterschließung, die den Zugriff auf heterogene, dezentral verteilte Informationsbestände und deren Verwaltung nach gemeinsamen Prinzipien ermöglichen sollen. Dabei geht CARMEN gezielt einen anderen Weg als die meisten bisherigen Ansätze in diesem Bereich, die versuchen, Homogenität und Konsistenz in einer dezentralen Informationslandschaft technikorientiert herzustellen, indem Verfahren entwickelt werden, durch die physikalisch auf verschiedene Dokumentenräume gleichzeitig zugegriffen werden kann. Eine rein technische Parallelisierung von Zugriffsmöglichkeiten reicht jedoch nicht aus, denn das Hauptproblem der inhaltlichen, strukturellen und konzeptionellen Differenz der einzelnen Datenbestände wird damit nicht gelöst. Um diese Differenzen zu kompensieren, werden Problemlösungen und Weiterentwicklungen innerhalb des Projekts CARMEN in drei Bereichen erarbeitet: (1) Metadaten (Dokumentbeschreibung, Retrieval, Verwaltung, Archivierung) (2) Methoden des Umgangs mit der verbleibenden Heterogenität der Datenbestände (3) Retrieval für strukturierte Dokumente mit Metadaten und heterogenen Datentypen. Diese drei Aufgabenbereiche hängen eng zusammen. Durch die Entwicklungen im Bereich der Metadaten soll einerseits die verlorengegangene Konsistenz partiell wiederhergestellt und auf eine den neuen Medien gerechte Basis gestellt werden. Andererseits sollen durch Verfahren zur Heterogenitätsbehandlung Dokumente mit unterschiedlicher Datenrelevanz und Inhaltserschließung aufeinander bezogen und retrievalseitig durch ein Rechercheverfahren erganzt werden, das den unterschiedlichen Datentypen gerecht wird Innerhalb des Gesamtprojekts CARMEN werden diese Aspekte arbeitsteilig behandelt. Acht Arbeitspakete (APs) befassen sich in Abstimmung miteinander mit je verschiedenen Schwerpunkten. Um die Koordination der Arbeiten der verschiedenen APs untereinander zu unterstützen, trafen sich die ca. 40 Projektbearbeiter am 1. und 2. Februar 2001 zum "CARMEN middle OfTheRoad Workshop" in Bonn. Anlässlich dieses Workshops wurden die inhaltlichen und technischen Ergebnisse, die in der ersten Hälfte der Projektlaufzeit von den einzelnen APs erzielt worden sind, in insgesamt 17 Präsentationen vorgestellt
  10. Avrahami, T.T.; Yau, L.; Si, L.; Callan, J.P.: ¬The FedLemur project : Federated search in the real world (2006) 0.03
    0.025751628 = product of:
      0.038627442 = sum of:
        0.02669988 = weight(_text_:retrieval in 5271) [ClassicSimilarity], result of:
          0.02669988 = score(doc=5271,freq=2.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.20052543 = fieldWeight in 5271, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=5271)
        0.011927563 = product of:
          0.035782687 = sum of:
            0.035782687 = weight(_text_:22 in 5271) [ClassicSimilarity], result of:
              0.035782687 = score(doc=5271,freq=2.0), product of:
                0.15414225 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04401763 = queryNorm
                0.23214069 = fieldWeight in 5271, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5271)
          0.33333334 = coord(1/3)
      0.6666667 = coord(2/3)
    
    Abstract
    Federated search and distributed information retrieval systems provide a single user interface for searching multiple full-text search engines. They have been an active area of research for more than a decade, but in spite of their success as a research topic, they are still rare in operational environments. This article discusses a prototype federated search system developed for the U.S. government's FedStats Web portal, and the issues addressed in adapting research solutions to this operational environment. A series of experiments explore how well prior research results, parameter settings, and heuristics apply in the FedStats environment. The article concludes with a set of lessons learned from this technology transfer effort, including observations about search engine quality in the real world.
    Date
    22. 7.2006 16:02:07
  11. Meiert, M.: Elektronische Publikationen an Hochschulen : Modellierung des elektronischen Publikationsprozesses am Beispiel der Universität Hildesheim (2006) 0.03
    0.025751628 = product of:
      0.038627442 = sum of:
        0.02669988 = weight(_text_:retrieval in 5974) [ClassicSimilarity], result of:
          0.02669988 = score(doc=5974,freq=2.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.20052543 = fieldWeight in 5974, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=5974)
        0.011927563 = product of:
          0.035782687 = sum of:
            0.035782687 = weight(_text_:22 in 5974) [ClassicSimilarity], result of:
              0.035782687 = score(doc=5974,freq=2.0), product of:
                0.15414225 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04401763 = queryNorm
                0.23214069 = fieldWeight in 5974, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5974)
          0.33333334 = coord(1/3)
      0.6666667 = coord(2/3)
    
    Date
    1. 9.2006 13:22:15
    Source
    Effektive Information Retrieval Verfahren in Theorie und Praxis: ausgewählte und erweiterte Beiträge des Vierten Hildesheimer Evaluierungs- und Retrievalworkshop (HIER 2005), Hildesheim, 20.7.2005. Hrsg.: T. Mandl u. C. Womser-Hacker
  12. Nicholson, D.; Steele, M.: CATRIONA : a distributed, locally-oriented, Z39.50 OPAC-based approach to cataloguing the Internet (1996) 0.03
    0.025751628 = product of:
      0.038627442 = sum of:
        0.02669988 = weight(_text_:retrieval in 603) [ClassicSimilarity], result of:
          0.02669988 = score(doc=603,freq=2.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.20052543 = fieldWeight in 603, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=603)
        0.011927563 = product of:
          0.035782687 = sum of:
            0.035782687 = weight(_text_:22 in 603) [ClassicSimilarity], result of:
              0.035782687 = score(doc=603,freq=2.0), product of:
                0.15414225 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04401763 = queryNorm
                0.23214069 = fieldWeight in 603, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=603)
          0.33333334 = coord(1/3)
      0.6666667 = coord(2/3)
    
    Abstract
    The aims of CATRIONA were: (1) to investigate the requirements for developing procedures and applications for cataloguing and retrieval of networked resources, and (2) to explore the feasibility of a collaborative project to develop such applications and procedures and integrate them with existing library systems. The project established that a distributed catalogue of networked resources integrated with standard Z39.50 library system OPAC interfaces with information on hard-copy resources is already a practical proposition at a basic level. At least one Z39.50 OPAC client can search remote Z39.50 OPACs, retrieve USMARC records with URLs in 856$u, load a viewer like Netscape, and use it to retrieve and display the remotely held electronic resource on the local workstation. A follow-up project on related issues is being finalised.
    Source
    Cataloging and classification quarterly. 22(1996) nos.3/4, S.127-141
  13. Callan, J.: Distributed information retrieval (2000) 0.03
    0.025433755 = product of:
      0.07630126 = sum of:
        0.07630126 = weight(_text_:retrieval in 31) [ClassicSimilarity], result of:
          0.07630126 = score(doc=31,freq=12.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.5730491 = fieldWeight in 31, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=31)
      0.33333334 = coord(1/3)
    
    Abstract
    A multi-database model of distributed information retrieval is presented, in which people are assumed to have access to many searchable text databases. In such an environment, full-text information retrieval consists of discovering database contents, ranking databases by their expected ability to satisfy the query, searching a small number of databases, and merging results returned by different databases. This paper presents algorithms for each task. It also discusses how to reorganize conventional test collections into multi-database testbeds, and evaluation methodologies for multi-database experiments. A broad and diverse group of experimental results is presented to demonstrate that the algorithms are effective, efficient, robust, and scalable
    Series
    The Kluwer international series on information retrieval; 7
    Source
    Advances in information retrieval: Recent research from the Center for Intelligent Information Retrieval. Ed.: W.B. Croft
  14. Croft, W.B.: Combining approaches to information retrieval (2000) 0.03
    0.02517289 = product of:
      0.07551867 = sum of:
        0.07551867 = weight(_text_:retrieval in 6862) [ClassicSimilarity], result of:
          0.07551867 = score(doc=6862,freq=16.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.5671716 = fieldWeight in 6862, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=6862)
      0.33333334 = coord(1/3)
    
    Abstract
    The combination of different text representations and search strategies has become a standard technique for improving the effectiveness of information retrieval. Combination, for example, has been studied extensively in the TREC evaluations and is the basis of the "meta-search" engines used on the Web. This paper examines the development of this technique, including both experimental results and the retrieval models that have been proposed as formal frameworks for combination. We show that combining approaches for information retrieval can be modeled as combining the outputs of multiple classifiers based on one or more representations, and that this simple model can provide explanations for many of the experimental results. We also show that this view of combination is very similar to the inference net model, and that a new approach to retrieval based on language models supports combination and can be integrated with the inference net model
    Series
    The Kluwer international series on information retrieval; 7
    Source
    Advances in information retrieval: Recent research from the Center for Intelligent Information Retrieval. Ed.: W.B. Croft
  15. López Vargas, M.A.: "Ilmenauer Verteiltes Information REtrieval System" (IVIRES) : eine neue Architektur zur Informationsfilterung in einem verteilten Information Retrieval System (2002) 0.03
    0.02517289 = product of:
      0.07551867 = sum of:
        0.07551867 = weight(_text_:retrieval in 4041) [ClassicSimilarity], result of:
          0.07551867 = score(doc=4041,freq=4.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.5671716 = fieldWeight in 4041, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.09375 = fieldNorm(doc=4041)
      0.33333334 = coord(1/3)
    
  16. Crestani, F.; Wu, S.: Testing the cluster hypothesis in distributed information retrieval (2006) 0.02
    0.024598189 = product of:
      0.073794566 = sum of:
        0.073794566 = weight(_text_:retrieval in 984) [ClassicSimilarity], result of:
          0.073794566 = score(doc=984,freq=22.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.554223 = fieldWeight in 984, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=984)
      0.33333334 = coord(1/3)
    
    Abstract
    How to merge and organise query results retrieved from different resources is one of the key issues in distributed information retrieval. Some previous research and experiments suggest that cluster-based document browsing is more effective than a single merged list. Cluster-based retrieval results presentation is based on the cluster hypothesis, which states that documents that cluster together have a similar relevance to a given query. However, while this hypothesis has been demonstrated to hold in classical information retrieval environments, it has never been fully tested in heterogeneous distributed information retrieval environments. Heterogeneous document representations, the presence of document duplicates, and disparate qualities of retrieval results, are major features of an heterogeneous distributed information retrieval environment that might disrupt the effectiveness of the cluster hypothesis. In this paper we report on an experimental investigation into the validity and effectiveness of the cluster hypothesis in highly heterogeneous distributed information retrieval environments. The results show that although clustering is affected by different retrieval results representations and quality, the cluster hypothesis still holds and that generating hierarchical clusters in highly heterogeneous distributed information retrieval environments is still a very effective way of presenting retrieval results to users.
  17. Fuhr, N.: Towards data abstraction in networked information retrieval systems (1999) 0.02
    0.020766575 = product of:
      0.06229972 = sum of:
        0.06229972 = weight(_text_:retrieval in 4517) [ClassicSimilarity], result of:
          0.06229972 = score(doc=4517,freq=2.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.46789268 = fieldWeight in 4517, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.109375 = fieldNorm(doc=4517)
      0.33333334 = coord(1/3)
    
  18. CARMEN : Content Analysis, Retrieval und Metadata: Effective Networking (1999) 0.02
    0.020766575 = product of:
      0.06229972 = sum of:
        0.06229972 = weight(_text_:retrieval in 5748) [ClassicSimilarity], result of:
          0.06229972 = score(doc=5748,freq=2.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.46789268 = fieldWeight in 5748, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.109375 = fieldNorm(doc=5748)
      0.33333334 = coord(1/3)
    
  19. Kunz, M.: Subject retrieval in distributed resources : a short review of recent developments (2003) 0.02
    0.017984381 = product of:
      0.05395314 = sum of:
        0.05395314 = weight(_text_:retrieval in 1624) [ClassicSimilarity], result of:
          0.05395314 = score(doc=1624,freq=6.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.40520695 = fieldWeight in 1624, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1624)
      0.33333334 = coord(1/3)
    
    Abstract
    Subject-based retrieval in distributed resources is a current problem in online searches for bibliographic references. Building portals to similar resources is only the ferst step, the subsequent navigation via different search interfaces presents certain difficulties. To make retrieval easier it is necessary to adapt these different resources. Potential approaches (standardisation as opposed to "cross-walks") and methods (automated as opposed to intellectual effort) will be discussed. This includes a Brief appraisal of the future of work with multilingual terminology: - The "classical" approach (Multilingual Thesauri), - The "Internet" approach (linking) Recent developments in mono- and multilingual environments will be presented (MACS, CARMEN, Economics Crosswalk).
  20. Hakala, J.: Z39.50-1995: information retrieval protocol : an introduction to the standard and it's usage (1996) 0.02
    0.016781926 = product of:
      0.050345775 = sum of:
        0.050345775 = weight(_text_:retrieval in 3340) [ClassicSimilarity], result of:
          0.050345775 = score(doc=3340,freq=4.0), product of:
            0.1331496 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.04401763 = queryNorm
            0.37811437 = fieldWeight in 3340, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=3340)
      0.33333334 = coord(1/3)
    
    Abstract
    This article describes the Internet information retrieval protocol, Z39.50, and it's usage. The services of Z39.50 are depicted, as are some important terms related to the standard. A description of the OPAC Network in Europe (ONE), an important Z39.50 implementation project is included

Languages

  • e 36
  • d 16
  • f 1
  • More… Less…

Types

  • a 44
  • el 5
  • m 4
  • x 4
  • r 1
  • s 1
  • More… Less…

Classifications