Search (19 results, page 1 of 1)

  • × language_ss:"e"
  • × theme_ss:"Information Gateway"
  • × type_ss:"el"
  1. Goodchild, M.F.: ¬The Alexandria Digital Library Project : review, assessment, and prospects (2004) 0.03
    0.031786613 = product of:
      0.06357323 = sum of:
        0.00823978 = product of:
          0.03295912 = sum of:
            0.03295912 = weight(_text_:based in 1153) [ClassicSimilarity], result of:
              0.03295912 = score(doc=1153,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.23302436 = fieldWeight in 1153, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1153)
          0.25 = coord(1/4)
        0.055333447 = product of:
          0.11066689 = sum of:
            0.11066689 = weight(_text_:assessment in 1153) [ClassicSimilarity], result of:
              0.11066689 = score(doc=1153,freq=2.0), product of:
                0.25917634 = queryWeight, product of:
                  5.52102 = idf(docFreq=480, maxDocs=44218)
                  0.04694356 = queryNorm
                0.4269946 = fieldWeight in 1153, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.52102 = idf(docFreq=480, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1153)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    The Alexandria Digital Library (ADL) was established in the late 1990s as a response to several perceived problems of traditional map libraries, notably access and organization. By 1999 it had evolved into an operational digital library, offering a well-defined set of services to a broad user community, based on an extensive collection of georeferenced information objects. The vision of ADL continues to evolve, as technology makes new services possible, as its users become more sophisticated and demanding, and as the broader field of geographic information science (GIScience) identifies new avenues for research and application.
  2. Summann, F.; Lossau, N.: Search engine technology and digital libraries : moving from theory to practice (2004) 0.03
    0.026661905 = product of:
      0.05332381 = sum of:
        0.008155267 = product of:
          0.032621067 = sum of:
            0.032621067 = weight(_text_:based in 1196) [ClassicSimilarity], result of:
              0.032621067 = score(doc=1196,freq=6.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.2306343 = fieldWeight in 1196, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1196)
          0.25 = coord(1/4)
        0.04516854 = weight(_text_:term in 1196) [ClassicSimilarity], result of:
          0.04516854 = score(doc=1196,freq=2.0), product of:
            0.21904005 = queryWeight, product of:
              4.66603 = idf(docFreq=1130, maxDocs=44218)
              0.04694356 = queryNorm
            0.20621133 = fieldWeight in 1196, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.66603 = idf(docFreq=1130, maxDocs=44218)
              0.03125 = fieldNorm(doc=1196)
      0.5 = coord(2/4)
    
    Abstract
    This article describes the journey from the conception of and vision for a modern search-engine-based search environment to its technological realisation. In doing so, it takes up the thread of an earlier article on this subject, this time from a technical viewpoint. As well as presenting the conceptual considerations of the initial stages, this article will principally elucidate the technological aspects of this journey. The starting point for the deliberations about development of an academic search engine was the experience we gained through the generally successful project "Digital Library NRW", in which from 1998 to 2000-with Bielefeld University Library in overall charge-we designed a system model for an Internet-based library portal with an improved academic search environment at its core. At the heart of this system was a metasearch with an availability function, to which we added a user interface integrating all relevant source material for study and research. The deficiencies of this approach were felt soon after the system was launched in June 2001. There were problems with the stability and performance of the database retrieval system, with the integration of full-text documents and Internet pages, and with acceptance by users, because users are increasingly performing the searches themselves using search engines rather than going to the library for help in doing searches. Since a long list of problems are also encountered using commercial search engines for academic use (in particular the retrieval of academic information and long-term availability), the idea was born for a search engine configured specifically for academic use. We also hoped that with one single access point founded on improved search engine technology, we could access the heterogeneous academic resources of subject-based bibliographic databases, catalogues, electronic newspapers, document servers and academic web pages.
  3. Bartolo, L.M.; Lowe, C.S.; Sadoway, D.R.; Powell, A.C.; Glotzer, S.C.: NSDL MatDL : exploring digital library roles (2005) 0.02
    0.022704724 = product of:
      0.04540945 = sum of:
        0.005885557 = product of:
          0.023542227 = sum of:
            0.023542227 = weight(_text_:based in 1181) [ClassicSimilarity], result of:
              0.023542227 = score(doc=1181,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.16644597 = fieldWeight in 1181, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1181)
          0.25 = coord(1/4)
        0.039523892 = product of:
          0.079047784 = sum of:
            0.079047784 = weight(_text_:assessment in 1181) [ClassicSimilarity], result of:
              0.079047784 = score(doc=1181,freq=2.0), product of:
                0.25917634 = queryWeight, product of:
                  5.52102 = idf(docFreq=480, maxDocs=44218)
                  0.04694356 = queryNorm
                0.30499613 = fieldWeight in 1181, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.52102 = idf(docFreq=480, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1181)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    A primary goal of the NSDL Materials Digital Library (MatDL) is to bring materials science research and education closer together. MatDL is exploring the various roles digital libraries can serve in the materials science community including: 1) supporting a virtual lab, 2) developing markup language applications, and 3) building tools for metadata capture. MatDL is being integrated into an MIT virtual laboratory experience. Early student self-assessment survey results expressed positive opinions of the potential value of MatDL in supporting a virtual lab and in accomplishing additional educational objectives. A separate survey suggested that the effectiveness of a virtual lab may approach that of a physical lab on some laboratory learning objectives. MatDL is collaboratively developing a materials property grapher (KSU and MIT) and a submission tool (KSU and U-M). MatML is an extensible markup language for exchanging materials information developed by materials data experts in industry, government, standards organizations, and professional societies. The web-based MatML grapher allows students to compare selected materials properties across approximately 80 MatML-tagged materials. The MatML grapher adds value in this educational context by allowing students to utilize real property data to make optimal material selection decisions. The submission tool has been integrated into the regular workflow of U-M students and researchers generating nanostructure images. It prompts users for domain-specific information, automatically generating and attaching keywords and editable descriptions.
  4. Kirriemuir, J.; Brickley, D.; Welsh, S.; Knight, J.; Hamilton, M.: Cross-searching subject gateways : the query routing and forward knowledge approach (1998) 0.00
    0.0036041527 = product of:
      0.014416611 = sum of:
        0.014416611 = product of:
          0.057666443 = sum of:
            0.057666443 = weight(_text_:based in 1252) [ClassicSimilarity], result of:
              0.057666443 = score(doc=1252,freq=12.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.4077077 = fieldWeight in 1252, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1252)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    A subject gateway, in the context of network-based resource access, can be defined as some facility that allows easier access to network-based resources in a defined subject area. The simplest types of subject gateways are sets of Web pages containing lists of links to resources. Some gateways index their lists of links and provide a simple search facility. More advanced gateways offer a much enhanced service via a system consisting of a resource database and various indexes, which can be searched and/or browsed through a Web-based interface. Each entry in the database contains information about a network-based resource, such as a Web page, Web site, mailing list or document. Entries are usually created by a cataloguer manually identifying a suitable resource, describing the resource using a template, and submitting the template to the database for indexing. Subject gateways are also known as subject-based information gateways (SBIGs), subject-based gateways, subject index gateways, virtual libraries, clearing houses, subject trees, pathfinders and other variations thereof. This paper describes the characteristics of some of the subject gateways currently accessible through the Web, and compares them to automatic "vacuum cleaner" type search engines, such as AltaVista. The application of WHOIS++, centroids, query routing, and forward knowledge to searching several of these subject gateways simultaneously is outlined. The paper concludes with looking at some of the issues facing subject gateway development in the near future. The paper touches on many of the issues mentioned in a previous paper in D-Lib Magazine, especially regarding resource-discovery related initiatives and services.
  5. Blosser, J.; Michaelson, R.; Routh. R.; Xia, P.: Defining the landscape of Web resources : Concluding Report of the BAER Web Resources Sub-Group (2000) 0.00
    0.003180102 = product of:
      0.012720408 = sum of:
        0.012720408 = product of:
          0.025440816 = sum of:
            0.025440816 = weight(_text_:22 in 1447) [ClassicSimilarity], result of:
              0.025440816 = score(doc=1447,freq=2.0), product of:
                0.16438834 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04694356 = queryNorm
                0.15476047 = fieldWeight in 1447, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1447)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    21. 4.2002 10:22:31
  6. Müller, B.; Poley, C.; Pössel, J.; Hagelstein, A.; Gübitz, T.: LIVIVO - the vertical search engine for life sciences (2017) 0.00
    0.0020808585 = product of:
      0.008323434 = sum of:
        0.008323434 = product of:
          0.033293735 = sum of:
            0.033293735 = weight(_text_:based in 3368) [ClassicSimilarity], result of:
              0.033293735 = score(doc=3368,freq=4.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.23539014 = fieldWeight in 3368, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3368)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    The explosive growth of literature and data in the life sciences challenges researchers to keep track of current advancements in their disciplines. Novel approaches in the life science like the One Health paradigm require integrated methodologies in order to link and connect heterogeneous information from databases and literature resources. Current publications in the life sciences are increasingly characterized by the employment of trans-disciplinary methodologies comprising molecular and cell biology, genetics, genomic, epigenomic, transcriptional and proteomic high throughput technologies with data from humans, plants, and animals. The literature search engine LIVIVO empowers retrieval functionality by incorporating various literature resources from medicine, health, environment, agriculture and nutrition. LIVIVO is developed in-house by ZB MED - Information Centre for Life Sciences. It provides a user-friendly and usability-tested search interface with a corpus of 55 Million citations derived from 50 databases. Standardized application programming interfaces are available for data export and high throughput retrieval. The search functions allow for semantic retrieval with filtering options based on life science entities. The service oriented architecture of LIVIVO uses four different implementation layers to deliver search services. A Knowledge Environment is developed by ZB MED to deal with the heterogeneity of data as an integrative approach to model, store, and link semantic concepts within literature resources and databases. Future work will focus on the exploitation of life science ontologies and on the employment of NLP technologies in order to improve query expansion, filters in faceted search, and concept based relevancy rankings in LIVIVO.
  7. Kriewel, S.; Klas, C.P.; Schaefer, A.; Fuhr, N.: DAFFODIL : strategic support for user-oriented access to heterogeneous digital libraries (2004) 0.00
    0.002059945 = product of:
      0.00823978 = sum of:
        0.00823978 = product of:
          0.03295912 = sum of:
            0.03295912 = weight(_text_:based in 4838) [ClassicSimilarity], result of:
              0.03295912 = score(doc=4838,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.23302436 = fieldWeight in 4838, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4838)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    DAFFODIL is a search system for digital libraries aiming at strategic support during the information search process. From a user point of view this strategic support is mainly implemented by high-level search functions, so-called stratagems, which provide functionality beyond today's digital libraries. Through the tight integration of stratagems and with the federation of heterogeneous digital libraries, DAFFODIL reaches high effects of synergy for information and services. These effects provide high-quality metadata for the searcher through an intuitively controllable user interface. The implementation of stratagems follows a tool-based model.
  8. Janée, G.; Frew, J.; Hill, L.L.: Issues in georeferenced digital libraries (2004) 0.00
    0.002059945 = product of:
      0.00823978 = sum of:
        0.00823978 = product of:
          0.03295912 = sum of:
            0.03295912 = weight(_text_:based in 1165) [ClassicSimilarity], result of:
              0.03295912 = score(doc=1165,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.23302436 = fieldWeight in 1165, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1165)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    Based on a decade's experience with the Alexandria Digital Library Project, seven issues are presented that arise in creating georeferenced digital libraries, and that appear to be intrinsic to the problem of creating any library-like information system that operates on georeferenced and geospatial resources. The first and foremost issue is providing discovery of georeferenced resources. Related to discovery are the issues of gazetteer integration and specialized ranking of search results. Strong data typing and scalability are implementation issues. Providing spatial context is a critical user interface issue. Finally, sophisticated resource access mechanisms are necessary to operate on geospatial resources.
  9. Urs, S.R.; Angrosh, M.A.: Ontology-based knowledge organization systems in digital libraries : a comparison of experiments in OWL and KAON ontologies (2006 (?)) 0.00
    0.0020388167 = product of:
      0.008155267 = sum of:
        0.008155267 = product of:
          0.032621067 = sum of:
            0.032621067 = weight(_text_:based in 2799) [ClassicSimilarity], result of:
              0.032621067 = score(doc=2799,freq=6.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.2306343 = fieldWeight in 2799, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2799)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    Grounded on a strong belief that ontologies enhance the performance of information retrieval systems, there has been an upsurge of interest in ontologies. Its importance is identified in diverse research fields such as knowledge engineering, knowledge representation, qualitative modeling, language engineering, database design, information integration, object-oriented analysis, information retrieval and extraction, knowledge management and agent-based systems design (Guarino, 1998). While the role-played by ontologies, automatically lends a place of legitimacy for these tools, research in this area gains greater significance in the wake of various challenges faced in the contemporary digital environment. With the objective of overcoming various pitfalls associated with current search mechanisms, ontologies are increasingly used for developing efficient information retrieval systems. An indicator of research interest in the area of ontology is the Swoogle, a search engine for Semantic Web documents, terms and data found on the Web (Ding, Li et al, 2004). Given the complex nature of the digital content archived in digital libraries, ontologies can be employed for designing efficient forms of information retrieval in digital libraries. Knowledge representation assumes greater significance due to its crucial role in ontology development. These systems aid in developing intelligent information systems, wherein the notion of intelligence implies the ability of the system to find implicit consequences of its explicitly represented knowledge (Baader and Nutt, 2003). Knowledge representation formalisms such as 'Description Logics' are used to obtain explicit knowledge representation of the subject domain. These representations are developed into ontologies, which are used for developing intelligent information systems. Against this backdrop, the paper examines the use of Description Logics for conceptually modeling a chosen domain, which would be utilized for developing domain ontologies. The knowledge representation languages identified for this purpose are Web Ontology Language (OWL) and KArlsruhe ONtology (KAON) language. Drawing upon the various technical constructs in developing ontology-based information systems, the paper explains the working of the prototypes and also presents a comparative study of the two prototypes.
  10. Arms, W.Y.; Blanchi, C.; Overly, E.A.: ¬An architecture for information in digital libraries (1997) 0.00
    0.0017839646 = product of:
      0.0071358583 = sum of:
        0.0071358583 = product of:
          0.028543433 = sum of:
            0.028543433 = weight(_text_:based in 1260) [ClassicSimilarity], result of:
              0.028543433 = score(doc=1260,freq=6.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.20180501 = fieldWeight in 1260, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1260)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    Flexible organization of information is one of the key design challenges in any digital library. For the past year, we have been working with members of the National Digital Library Project (NDLP) at the Library of Congress to build an experimental system to organize and store library collections. This is a report on the work. In particular, we describe how a few technical building blocks are used to organize the material in collections, such as the NDLP's, and how these methods fit into a general distributed computing framework. The technical building blocks are part of a framework that evolved as part of the Computer Science Technical Reports Project (CSTR). This framework is described in the paper, "A Framework for Distributed Digital Object Services", by Robert Kahn and Robert Wilensky (1995). The main building blocks are: "digital objects", which are used to manage digital material in a networked environment; "handles", which identify digital objects and other network resources; and "repositories", in which digital objects are stored. These concepts are amplified in "Key Concepts in the Architecture of the Digital Library", by William Y. Arms (1995). In summer 1995, after earlier experimental development, work began on the implementation of a full digital library system based on this framework. In addition to Kahn/Wilensky and Arms, several working papers further elaborate on the design concepts. A paper by Carl Lagoze and David Ely, "Implementation Issues in an Open Architectural Framework for Digital Object Services", delves into some of the repository concepts. The initial repository implementation was based on a paper by Carl Lagoze, Robert McGrath, Ed Overly and Nancy Yeager, "A Design for Inter-Operable Secure Object Stores (ISOS)". Work on the handle system, which began in 1992, is described in a series of papers that can be found on the Handle Home Page. The National Digital Library Program (NDLP) at the Library of Congress is a large scale project to convert historic collections to digital form and make them widely available over the Internet. The program is described in two articles by Caroline R. Arms, "Historical Collections for the National Digital Library". The NDLP itself draws on experience gained through the earlier American Memory Program. Based on this work, we have built a pilot system that demonstrates how digital objects can be used to organize complex materials, such as those found in the NDLP. The pilot was demonstrated to members of the library in July 1996. The pilot system includes the handle system for identifying digital objects, a pilot repository to store them, and two user interfaces: one designed for librarians to manage digital objects in the repository, the other for library patrons to access the materials stored in the repository. Materials from the NDLP's Coolidge Consumerism compilation have been deposited into the pilot repository. They include a variety of photographs and texts, converted to digital form. The pilot demonstrates the use of handles for identifying such material, the use of meta-objects for managing sets of digital objects, and the choice of metadata. We are now implementing an enhanced prototype system for completion in early 1997.
  11. Doerr, M.; Gradmann, S.; Hennicke, S.; Isaac, A.; Meghini, C.; Van de Sompel, H.: ¬The Europeana Data Model (EDM) (2010) 0.00
    0.0017656671 = product of:
      0.0070626684 = sum of:
        0.0070626684 = product of:
          0.028250674 = sum of:
            0.028250674 = weight(_text_:based in 3967) [ClassicSimilarity], result of:
              0.028250674 = score(doc=3967,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.19973516 = fieldWeight in 3967, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3967)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    The Europeana Data Model (EDM) is a new approach towards structuring and representing data delivered to Europeana by the various contributing cultural heritage institutions. The model aims at greater expressivity and flexibility in comparison to the current Europeana Semantic Elements (ESE), which it is destined to replace. The design principles underlying the EDM are based on the core principles and best practices of the Semantic Web and Linked Data efforts to which Europeana wants to contribute. The model itself builds upon established standards like RDF(S), OAI-ORE, SKOS, and Dublin Core. It acts as a common top-level ontology which retains original data models and information perspectives while at the same time enabling interoperability. The paper elaborates on the aforementioned aspects and the design principles which drove the development of the EDM.
  12. Severiens, T.; Hohlfeld, M.; Zimmermann, K.; Hilf, E.R.: PhysDoc - a distributed network of physics institutions documents : collecting, indexing, and searching high quality documents by using harvest (2000) 0.00
    0.0014713892 = product of:
      0.005885557 = sum of:
        0.005885557 = product of:
          0.023542227 = sum of:
            0.023542227 = weight(_text_:based in 6470) [ClassicSimilarity], result of:
              0.023542227 = score(doc=6470,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.16644597 = fieldWeight in 6470, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=6470)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    PhysNet offers online services that enable a physicist to keep in touch with the worldwide physics community and to receive all information he or she may need. In addition to being of great value to physicists, these services are practical examples of the use of modern methods of digital libraries, in particular the use of metadata harvesting. One service is PhysDoc. This consists of a Harvest-based online information broker- and gatherer-network, which harvests information from the local web-servers of professional physics institutions worldwide (mostly in Europe and USA so far). PhysDoc focuses on scientific information posted by the individual scientist at his local server, such as documents, publications, reports, publication lists, and lists of links to documents. All rights are reserved for the authors who are responsible for the content and quality of their documents. PhysDis is an analogous service but specifically for university theses, with their dual requirements of examination work and publication. The strategy is to select high quality sites containing metadata. We report here on the present status of PhysNet, our experience in operating it, and the development of its usage. To continuously involve authors, research groups, and national societies is considered crucial for a future stable service.
  13. Fang, L.: ¬A developing search service : heterogeneous resources integration and retrieval system (2004) 0.00
    0.0014713892 = product of:
      0.005885557 = sum of:
        0.005885557 = product of:
          0.023542227 = sum of:
            0.023542227 = weight(_text_:based in 1193) [ClassicSimilarity], result of:
              0.023542227 = score(doc=1193,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.16644597 = fieldWeight in 1193, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1193)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    This article describes two approaches for searching heterogeneous resources, which are explained as they are used in two corresponding existing systems-RIRS (Resource Integration Retrieval System) and HRUSP (Heterogeneous Resource Union Search Platform). On analyzing the existing systems, a possible framework-the MUSP (Multimetadata-Based Union Search Platform) is presented. Libraries now face a dilemma. On one hand, libraries subscribe to many types of database retrieval systems that are produced by various providers. The libraries build their data and information systems independently. This results in highly heterogeneous and distributed systems at the technical level (e.g., different operating systems and user interfaces) and at the conceptual level (e.g., the same objects are named using different terms). On the other hand, end users want to access all these heterogeneous data via a union interface, without having to know the structure of each information system or the different retrieval methods used by the systems. Libraries must achieve a harmony between information providers and users. In order to bridge the gap between the service providers and the users, it would seem that all source databases would need to be rebuilt according to a uniform data structure and query language, but this seems impossible. Fortunately, however, libraries and information and technology providers are now making an effort to find a middle course that meets the requirements of both data providers and users. They are doing this through resource integration.
  14. Spink, A.; Wilson, T.; Ellis, D.; Ford, N.: Modeling users' successive searches in digital environments : a National Science Foundation/British Library funded study (1998) 0.00
    0.0014566011 = product of:
      0.0058264043 = sum of:
        0.0058264043 = product of:
          0.023305617 = sum of:
            0.023305617 = weight(_text_:based in 1255) [ClassicSimilarity], result of:
              0.023305617 = score(doc=1255,freq=4.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.1647731 = fieldWeight in 1255, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1255)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    As digital libraries become a major source of information for many people, we need to know more about how people seek and retrieve information in digital environments. Quite commonly, users with a problem-at-hand and associated question-in-mind repeatedly search a literature for answers, and seek information in stages over extended periods from a variety of digital information resources. The process of repeatedly searching over time in relation to a specific, but possibly an evolving information problem (including changes or shifts in a variety of variables), is called the successive search phenomenon. The study outlined in this paper is currently investigating this new and little explored line of inquiry for information retrieval, Web searching, and digital libraries. The purpose of the research project is to investigate the nature, manifestations, and behavior of successive searching by users in digital environments, and to derive criteria for use in the design of information retrieval interfaces and systems supporting successive searching behavior. This study includes two related projects. The first project is based in the School of Library and Information Sciences at the University of North Texas and is funded by a National Science Foundation POWRE Grant <http://www.nsf.gov/cgi-bin/show?award=9753277>. The second project is based at the Department of Information Studies at the University of Sheffield (UK) and is funded by a grant from the British Library <http://www.shef. ac.uk/~is/research/imrg/uncerty.html> Research and Innovation Center. The broad objectives of each project are to examine the nature and extent of successive search episodes in digital environments by real users over time. The specific aim of the current project is twofold: * To characterize progressive changes and shifts that occur in: user situational context; user information problem; uncertainty reduction; user cognitive styles; cognitive and affective states of the user, and consequently in their queries; and * To characterize related changes over time in the type and use of information resources and search strategies particularly related to given capabilities of IR systems, and IR search engines, and examine changes in users' relevance judgments and criteria, and characterize their differences. The study is an observational, longitudinal data collection in the U.S. and U.K. Three questionnaires are used to collect data: reference, client post search and searcher post search questionnaires. Each successive search episode with a search intermediary for textual materials on the DIALOG Information Service is audiotaped and search transaction logs are recorded. Quantitative analysis includes statistical analysis using Likert scale data from the questionnaires and log-linear analysis of sequential data. Qualitative methods include: content analysis, structuring taxonomies; and diagrams to describe shifts and transitions within and between each search episode. Outcomes of the study are the development of appropriate model(s) for IR interactions in successive search episodes and the derivation of a set of design criteria for interfaces and systems supporting successive searching.
  15. Buckland, M.; Lancaster, L.: Combining place, time, and topic : the Electronic Cultural Atlas Initiative (2004) 0.00
    0.0011771114 = product of:
      0.0047084456 = sum of:
        0.0047084456 = product of:
          0.018833783 = sum of:
            0.018833783 = weight(_text_:based in 1194) [ClassicSimilarity], result of:
              0.018833783 = score(doc=1194,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.13315678 = fieldWeight in 1194, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1194)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    The Electronic Cultural Atlas Initiative was formed to encourage scholarly communication and the sharing of data among researchers who emphasize the relationships between place, time, and topic in the study of culture and history. In an effort to develop better tools and practices, The Electronic Cultural Atlas Initiative has sponsored the collaborative development of software for downloading and editing geo-temporal data to create dynamic maps, a clearinghouse of shared datasets accessible through a map-based interface, projects on format and content standards for gazetteers and time period directories, studies to improve geo-temporal aspects in online catalogs, good practice guidelines for preparing e-publications with dynamic geo-temporal displays, and numerous international conferences. The Electronic Cultural Atlas Initiative (ECAI) grew out of discussions among an international group of scholars interested in religious history and area studies. It was established as a unit under the Dean of International and Area Studies at the University of California, Berkeley in 1997. ECAI's mission is to promote an international collaborative effort to transform humanities scholarship through use of the digital environment to share data and by placing greater emphasis on the notions of place and time. Professor Lewis Lancaster is the Director. Professor Michael Buckland, with a library and information studies background, joined the effort as Co-Director in 2000. Assistance from the Lilly Foundation, the California Digital Library (University of California), and other sources has enabled ECAI to nurture a community; to develop a catalog ("clearinghouse") of Internet-accessible georeferenced resources; to support the development of software for obtaining, editing, manipulating, and dynamically visualizing geo-temporally encoded data; and to undertake research and development projects as needs and resources determine. Several hundred scholars worldwide, from a wide range of disciplines, are informally affiliated with ECAI, all interested in shared use of historical and cultural data. The Academia Sinica (Taiwan), The British Library, and the Arts and Humanities Data Service (UK) are among the well-known affiliates. However, ECAI mainly comprises individual scholars and small teams working on their own small projects on a very wide range of cultural, social, and historical topics. Numerous specialist committees have been fostering standardization and collaboration by area and by themes such as trade-routes, cities, religion, and sacred sites.
  16. Zia, L.L.: Growing a national learning environments and resources network for science, mathematics, engineering, and technology education : current issues and opportunities for the NSDL program (2001) 0.00
    0.0011771114 = product of:
      0.0047084456 = sum of:
        0.0047084456 = product of:
          0.018833783 = sum of:
            0.018833783 = weight(_text_:based in 1217) [ClassicSimilarity], result of:
              0.018833783 = score(doc=1217,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.13315678 = fieldWeight in 1217, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1217)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    The National Science Foundation's (NSF) National Science, Mathematics, Engineering, and Technology Education Digital Library (NSDL) program seeks to create, develop, and sustain a national digital library supporting science, mathematics, engineering, and technology (SMET) education at all levels -- preK-12, undergraduate, graduate, and life-long learning. The resulting virtual institution is expected to catalyze and support continual improvements in the quality of science, mathematics, engineering, and technology (SMET) education in both formal and informal settings. The vision for this program has been explored through a series of workshops over the past several years and documented in accompanying reports and monographs. (See [1-7, 10, 12, and 13].) These efforts have led to a characterization of the digital library as a learning environments and resources network for science, mathematics, engineering, and technology education, that is: * designed to meet the needs of learners, in both individual and collaborative settings; * constructed to enable dynamic use of a broad array of materials for learning primarily in digital format; and * managed actively to promote reliable anytime, anywhere access to quality collections and services, available both within and without the network. Underlying the NSDL program are several working assumptions. First, while there is currently no lack of "great piles of content" on the Web, there is an urgent need for "piles of great content". The difficulties in discovering and verifying the authority of appropriate Web-based material are certainly well known, yet there are many examples of learning resources of great promise available (particularly those exploiting the power of multiple media), with more added every day. The breadth and interconnectedness of the Web are simultaneously a great strength and shortcoming. Second, the "unit" or granularity of educational content can and will shrink, affording the opportunity for users to become creators and vice versa, as learning objects are reused, repackaged, and repurposed. To be sure, this scenario cannot take place without serious attention to intellectual property and digital rights management concerns. But new models and technologies are being explored (see a number of recent articles in the January issue of D-Lib Magazine). Third, there is a need for an "organizational infrastructure" that facilitates connections between distributed users and distributed content, as alluded to in the third bullet above. Finally, while much of the ongoing use of the library is envisioned to be "free" in the sense of the public good, there is an opportunity and a need to consider multiple alternative models of sustainability, particularly in the area of services offered by the digital library. More details about the NSDL program including information about proposal deadlines and current awards may be found at <http://www.ehr.nsf.gov/ehr/due/programs/nsdl>.
  17. Rusch-Feja, D.; Becker, H.J.: Global Info : the German digital libraries project (1999) 0.00
    0.0011771114 = product of:
      0.0047084456 = sum of:
        0.0047084456 = product of:
          0.018833783 = sum of:
            0.018833783 = weight(_text_:based in 1242) [ClassicSimilarity], result of:
              0.018833783 = score(doc=1242,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.13315678 = fieldWeight in 1242, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1242)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    The concept for the German Digital Libraries Program is imbedded in the Information Infrastructure Program of the German Federal Government for the years 1996-2000 which has been explicated in the Program Paper entitled "Information as Raw Material for Innovation".3 The Program Paper was published 1996 by the Federal Ministry for Education, Research, and Technology. The actual grants program "Global Info" was initiated by the Information and Communication Commission of the Joint Learned Societies to further technological advancement in enabling all researchers in Germany direct access to literature, research results, and other relevant information. This Commission was founded by four of the learned societies in 1995, and it has sponsored a series of workshops to increase awareness of leading edge technology and innovations in accessing electronic information sources. Now, nine of the leading research-level learned societies -- often those with umbrella responsibilities for other learned societies in their field -- are members of the Information and Communication Commission and represent the mathematicians, physicists, computer scientists, chemists, educational researchers, sociologists, psychologists, biologists and information technologists in the German Association of Engineers. (The German professional librarian societies are not members, as such, of this Commission, but are represented through delegates from libraries in the learned societies and in the future, hopefully, also by the German Association of Documentalists or through the cooperation between the documentalist and librarian professional societies.) The Federal Ministry earmarked 60 Million German Marks for projects within the framework of the German Digital Libraries Program in two phases over the next six years. The scope for the German Digital Libraries Program was announced in a press release in April 1997,4 and the first call for preliminary projects and expressions of interest in participation ended in July 1997. The Consortium members were suggested by the Information and Communication Commission of the Learned Societies (IuK Kommission), by key scientific research funding agencies in the German government, and by the publishers themselves. The first official meeting of the participants took place on December 1, 1997, at the Deutsche Bibliothek, located in the renowned center of German book trade, Frankfurt, thus documenting the active role and participation of libraries and publishers. In contrast to the Digital Libraries Project of the National Science Foundation in the United States, the German Digital Libraries project is based on furthering cooperation with universities, scientific publishing houses (including various international publishers), book dealers, and special subject information centers, as well as academic and research libraries. The goals of the German Digital Libraries Project are to achieve: 1) efficient access to world wide information; 2) directly from the scientist's desktop; 3) while providing the organization for and stimulating fundamental structural changes in the information and communication process of the scientific community.
  18. Niggemann, E.: Europeana: connecting cultural heritage (2009) 0.00
    0.0011771114 = product of:
      0.0047084456 = sum of:
        0.0047084456 = product of:
          0.018833783 = sum of:
            0.018833783 = weight(_text_:based in 2816) [ClassicSimilarity], result of:
              0.018833783 = score(doc=2816,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.13315678 = fieldWeight in 2816, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2816)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    The European Commission's goal for Europeana is to make European information resources easier to use in an online environment. It will build on Europe's rich heritage, combining multicultural and multilingual environments with technological advances and new business models. The Europeana prototype is the result of a 2-year project that began in July 2007. Europeana.eu went live on 20 November 2008, launched by Viviane Reding, European Commissioner for Information Society and Media. Europeana.eu is about ideas and inspiration. It links the user to 2 million digital items: images, text, sounds and videos. Europeana is a Thematic Network funded by the European Commission under the eContentplus programme, as part of the i2010 policy. Originally known as the European digital library network - EDLnet - it is a partnership of 100 representatives of heritage and knowledge organisations and IT experts from throughout Europe. They contribute to the work packages that are solving the technical and usability issues. The project is run by a core team based in the national library of the Netherlands, the Koninklijke Bibliotheek. It builds on the project management and technical expertise developed by The European Library, which is a service of the Conference of European National Librarians. Content is added via so called aggregators, national or domain specific portals aggegrating digital content and channelling it to Europeana. Most of these portals are being developed in the framework of EU funded projects, e.g. European Film Gateway, Athena and EuropeanaLocal. Overseeing the project is the EDL Foundation, which includes key European cultural heritage associations from the four domains. The Foundation's statutes commit members to: * providing access to Europe's cultural and scientific heritage through a cross-domain portal; * co-operating in the delivery and sustainability of the joint portal; * stimulating initiatives to bring together existing digital content; * supporting digitisation of Europe's cultural and scientific heritage. Europeana.eu is a prototype. Europeana Version 1.0 is being developed and will be launched in 2010 with links to over 6 million digital objects.
  19. Heery, R.; Carpenter, L.; Day, M.: Renardus project developments and the wider digital library context (2001) 0.00
    0.0010404292 = product of:
      0.004161717 = sum of:
        0.004161717 = product of:
          0.016646868 = sum of:
            0.016646868 = weight(_text_:based in 1219) [ClassicSimilarity], result of:
              0.016646868 = score(doc=1219,freq=4.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.11769507 = fieldWeight in 1219, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=1219)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    For those building digital library services, the organisational barriers are sometimes far more intractable than technological issues. This was firmly flagged in one of the first workshops focusing specifically on the digital library research agenda: Digital libraries are not simply technological constructs; they exist within a rich legal, social, and economic context, and will succeed only to the extent that they meet these broader needs. The innovatory drive within the development of digital library services thrives on the tension between meeting both technical and social imperatives. The Renardus project partners have previously taken parts in projects establishing the technical basis for subject gateways (e.g., ROADS, DESIRE], EELS) and are aware that technical barriers to interoperability are outweighed by challenges relating to the organisational and business models used. Within the Renardus project there has been a determination to address these organisational and business issues from the beginning. Renardus intends initially to create a pilot service, targeting the European scholar with a single point of access to quality selected Web resources. Looking ahead beyond current project funding, it aims to create the organisational and technological infrastructure for a sustainable service. This means the project is concerned with the range of processes required to establish a viable service, and is explicitly addressing business issues as well as providing a technical infrastructure. The overall aim of Renardus is to establish a collaborative framework for European subject gateways that will benefit both users in terms of enhanced services, and the gateways themselves in terms of shared solutions. In order to achieve this aim, Renardus will provide firstly a pilot service for the European academic and research communities brokering access to those European-based information gateways that currently participate in the project; in other words, brokering to gateways that are already in existence. Secondly the project will explore ways to establish the organisational basis for co-operative efforts such as metadata sharing, joint technical solutions and agreement on standardisation. It is intended that this exploration will feed back valuable experience to the individual participating gateways to suggest ways their services can be enhanced.
    The Renardus project has brought together gateways that are 'large-scale national initiatives'. Within the European context this immediately introduces a diversity of organisations, as responsibility for national gateway initiatives is located differently, for example, in national libraries, national agencies with responsibility for educational technology infrastructure, and within universities or consortia of universities. Within the project, gateways are in some cases represented directly by their own personnel, in some cases by other departments or research centres, but not always by the people responsible for providing the gateway service. For example, the UK Resource Discovery Network (RDN) is represented in the project by UKOLN (formerly part of the Resource Discovery Network Centre) and the Institute of Learning and Research Technology (ILRT), University of Bristol -- an RDN 'hub' service provider -- who are primarily responsible for dissemination. Since the start of the project there have been changes within the organisational structures providing gateways and within the service ambitions of gateways themselves. Such lack of stability is inherent within the Internet service environment, and this presents challenges to Renardus activity that has to be planned for a three-year period. For example, within the gateway's funding environment there is now an exploration of 'subject portals' offering more extended services than gateways. There is also potential commercial interest for including gateways as a value-added component to existing commercial services, and new offerings from possible competitors such as Google's Web Directory and country based services. This short update on the Renardus project intends to inform the reader of progress within the project and to give some wider context to its main themes by locating the project within the broader arena of digital library activity. There are twelve partners in the project from Denmark, Finland, France, Germany, the Netherlands and Sweden, as well as the UK. In particular we will focus on the specific activity in which UKOLN is involved: the architectural design, the specification of functional requirements, reaching consensus on a collaborative business model, etc. We will also consider issues of metadata management where all partners have interests. We will highlight implementation issues that connect to areas of debate elsewhere. In particular we see connections with activity related to establishing architectural models for digital library services, connections to the services that may emerge from metadata sharing using the Open Archives Initiative metadata sharing protocol, and links with work elsewhere on navigation of digital information spaces by means of controlled vocabularies.