Search (439 results, page 22 of 22)

  • × type_ss:"a"
  • × type_ss:"el"
  1. Donahue, J.; Hendricks, L.A.; Guadarrama, S.; Rohrbach, M.; Venugopalan, S.; Saenko, K.; Darrell, T.: Long-term recurrent convolutional networks for visual recognition and description (2014) 0.00
    0.0028656456 = product of:
      0.008596936 = sum of:
        0.008596936 = product of:
          0.025790809 = sum of:
            0.025790809 = weight(_text_:retrieval in 1873) [ClassicSimilarity], result of:
              0.025790809 = score(doc=1873,freq=2.0), product of:
                0.15433937 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.051022716 = queryNorm
                0.16710453 = fieldWeight in 1873, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1873)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    Models based on deep convolutional networks have dominated recent image interpretation tasks; we investigate whether models which are also recurrent, or "temporally deep", are effective for tasks involving sequences, visual and otherwise. We develop a novel recurrent convolutional architecture suitable for large-scale visual learning which is end-to-end trainable, and demonstrate the value of these models on benchmark video recognition tasks, image description and retrieval problems, and video narration challenges. In contrast to current models which assume a fixed spatio-temporal receptive field or simple temporal averaging for sequential processing, recurrent convolutional models are "doubly deep" in that they can be compositional in spatial and temporal "layers". Such models may have advantages when target concepts are complex and/or training data are limited. Learning long-term dependencies is possible when nonlinearities are incorporated into the network state updates. Long-term RNN models are appealing in that they directly can map variable-length inputs (e.g., video frames) to variable length outputs (e.g., natural language text) and can model complex temporal dynamics; yet they can be optimized with backpropagation. Our recurrent long-term models are directly connected to modern visual convnet models and can be jointly trained to simultaneously learn temporal dynamics and convolutional perceptual representations. Our results show such models have distinct advantages over state-of-the-art models for recognition or generation which are separately defined and/or optimized.
  2. Cao, N.; Sun, J.; Lin, Y.-R.; Gotz, D.; Liu, S.; Qu, H.: FacetAtlas : Multifaceted visualization for rich text corpora (2010) 0.00
    0.0028656456 = product of:
      0.008596936 = sum of:
        0.008596936 = product of:
          0.025790809 = sum of:
            0.025790809 = weight(_text_:retrieval in 3366) [ClassicSimilarity], result of:
              0.025790809 = score(doc=3366,freq=2.0), product of:
                0.15433937 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.051022716 = queryNorm
                0.16710453 = fieldWeight in 3366, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3366)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
  3. Smiraglia, R.P.: Facets as discourse in knowledge organization : a case study in LISTA (2017) 0.00
    0.0028656456 = product of:
      0.008596936 = sum of:
        0.008596936 = product of:
          0.025790809 = sum of:
            0.025790809 = weight(_text_:retrieval in 3855) [ClassicSimilarity], result of:
              0.025790809 = score(doc=3855,freq=2.0), product of:
                0.15433937 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.051022716 = queryNorm
                0.16710453 = fieldWeight in 3855, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3855)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    Knowledge Organization Systems (KOSs) use arrays of related concepts to capture the ontological content of a domain; hierarchical structures are typical of such systems. Some KOSs also employ sets of crossconceptual descriptors that express different dimensions within a domain-facets. The recent increase in the prominence of facets and faceted systems has had major impact on the intension of the KO domain and this is visible in the domain's literature. An interesting question is how the discourse surrounding facets in KO and in related domains such as information science might be described. The present paper reports one case study in an ongoing research project to investigate the discourse of facets in KO. In this particular case, the formal current research literature represented by inclusion in the "Library, Information Science & Technology Abstracts, Full Text" (LISTA) database is analyzed to discover aspects of the research front and its ongoing discourse concerning facets. A datasets of 1682 citations was analyzed. Results show thinking concerning information retrieval and the semantic web resides alongside implementation of faceted searching and the growth of faceted thesauri. Faceted classification remains important to the discourse, but the use of facet analysis is linked directly to applied aspects of information science.
  4. Spink, A.; Wilson, T.; Ellis, D.; Ford, N.: Modeling users' successive searches in digital environments : a National Science Foundation/British Library funded study (1998) 0.00
    0.002836844 = product of:
      0.008510532 = sum of:
        0.008510532 = product of:
          0.025531596 = sum of:
            0.025531596 = weight(_text_:retrieval in 1255) [ClassicSimilarity], result of:
              0.025531596 = score(doc=1255,freq=4.0), product of:
                0.15433937 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.051022716 = queryNorm
                0.16542503 = fieldWeight in 1255, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1255)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    As digital libraries become a major source of information for many people, we need to know more about how people seek and retrieve information in digital environments. Quite commonly, users with a problem-at-hand and associated question-in-mind repeatedly search a literature for answers, and seek information in stages over extended periods from a variety of digital information resources. The process of repeatedly searching over time in relation to a specific, but possibly an evolving information problem (including changes or shifts in a variety of variables), is called the successive search phenomenon. The study outlined in this paper is currently investigating this new and little explored line of inquiry for information retrieval, Web searching, and digital libraries. The purpose of the research project is to investigate the nature, manifestations, and behavior of successive searching by users in digital environments, and to derive criteria for use in the design of information retrieval interfaces and systems supporting successive searching behavior. This study includes two related projects. The first project is based in the School of Library and Information Sciences at the University of North Texas and is funded by a National Science Foundation POWRE Grant <http://www.nsf.gov/cgi-bin/show?award=9753277>. The second project is based at the Department of Information Studies at the University of Sheffield (UK) and is funded by a grant from the British Library <http://www.shef. ac.uk/~is/research/imrg/uncerty.html> Research and Innovation Center. The broad objectives of each project are to examine the nature and extent of successive search episodes in digital environments by real users over time. The specific aim of the current project is twofold: * To characterize progressive changes and shifts that occur in: user situational context; user information problem; uncertainty reduction; user cognitive styles; cognitive and affective states of the user, and consequently in their queries; and * To characterize related changes over time in the type and use of information resources and search strategies particularly related to given capabilities of IR systems, and IR search engines, and examine changes in users' relevance judgments and criteria, and characterize their differences. The study is an observational, longitudinal data collection in the U.S. and U.K. Three questionnaires are used to collect data: reference, client post search and searcher post search questionnaires. Each successive search episode with a search intermediary for textual materials on the DIALOG Information Service is audiotaped and search transaction logs are recorded. Quantitative analysis includes statistical analysis using Likert scale data from the questionnaires and log-linear analysis of sequential data. Qualitative methods include: content analysis, structuring taxonomies; and diagrams to describe shifts and transitions within and between each search episode. Outcomes of the study are the development of appropriate model(s) for IR interactions in successive search episodes and the derivation of a set of design criteria for interfaces and systems supporting successive searching.
  5. Dushay, N.: Visualizing bibliographic metadata : a virtual (book) spine viewer (2004) 0.00
    0.0024476624 = product of:
      0.0073429868 = sum of:
        0.0073429868 = product of:
          0.02202896 = sum of:
            0.02202896 = weight(_text_:online in 1197) [ClassicSimilarity], result of:
              0.02202896 = score(doc=1197,freq=4.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.142261 = fieldWeight in 1197, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1197)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    User interfaces for digital information discovery often require users to click around and read a lot of text in order to find the text they want to read-a process that is often frustrating and tedious. This is exacerbated because of the limited amount of text that can be displayed on a computer screen. To improve the user experience of computer mediated information discovery, information visualization techniques are applied to the digital library context, while retaining traditional information organization concepts. In this article, the "virtual (book) spine" and the virtual spine viewer are introduced. The virtual spine viewer is an application which allows users to visually explore large information spaces or collections while also allowing users to hone in on individual resources of interest. The virtual spine viewer introduced here is an alpha prototype, presented to promote discussion and further work. Information discovery changed radically with the introduction of computerized library access catalogs, the World Wide Web and its search engines, and online bookstores. Yet few instances of these technologies provide a user experience analogous to walking among well-organized, well-stocked bookshelves-which many people find useful as well as pleasurable. To put it another way, many of us have heard or voiced complaints about the paucity of "online browsing"-but what does this really mean? In traditional information spaces such as libraries, often we can move freely among the books and other resources. When we walk among organized, labeled bookshelves, we get a sense of the information space-we take in clues, perhaps unconsciously, as to the scope of the collection, the currency of resources, the frequency of their use, etc. We also enjoy unexpected discoveries such as finding an interesting resource because library staff deliberately located it near similar resources, or because it was miss-shelved, or because we saw it on a bookshelf on the way to the water fountain.
  6. DiLauro, T.; Choudhury, G.S.; Patton, M.; Warner, J.W.; Brown, E.W.: Automated name authority control and enhanced searching in the Levy collection (2001) 0.00
    0.002307678 = product of:
      0.006923034 = sum of:
        0.006923034 = product of:
          0.0207691 = sum of:
            0.0207691 = weight(_text_:online in 1160) [ClassicSimilarity], result of:
              0.0207691 = score(doc=1160,freq=2.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.13412495 = fieldWeight in 1160, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1160)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    This paper is the second in a series in D-Lib Magazine and describes a workflow management system being developed by the Digital Knowledge Center (DKC) at the Milton S. Eisenhower Library (MSEL) of The Johns Hopkins University. Based on experience from digitizing the Lester S. Levy Collection of Sheet Music, it was apparent that large-scale digitization efforts require a significant amount of human labor that is both time-consuming and costly. Consequently, this workflow management system aims to reduce the amount of human labor and time for large-scale digitization projects. The mission of this second phase of the project ("Levy II") can be summarized as follows: * Reduce costs for large collection ingestion by creating a suite of open-source processes, tools, and interfaces for workflow management * Increase access capabilities by providing a suite of research tools * Demonstrate utility of tools and processes with a subset of the online Levy Collection The cornerstones of the workflow management system include optical music recognition (OMR) software and an automated name authority control system (ANAC). The OMR software generates a logical representation of the score for sound generation, music searching, and musicological research. The ANAC disambiguates names, associating each name with an individual (e.g., the composer Septimus Winner also published under the pseudonyms Alice Hawthorne and Apsley Street, among others). Complementing the workflow tools, a suite of research tools focuses upon enhanced searching capabilities through the development and application of a fast, disk-based search engine for lyrics and music and the incorporation of an XML structure for metadata. The first paper (Choudhury et al. 2001) described the OMR software and musical components of Levy II. This paper focuses on the metadata and intellectual access components that include automated name authority control and the aforementioned search engine.
  7. Cohen, D.J.: From Babel to knowledge : data mining large digital collections (2006) 0.00
    0.002307678 = product of:
      0.006923034 = sum of:
        0.006923034 = product of:
          0.0207691 = sum of:
            0.0207691 = weight(_text_:online in 1178) [ClassicSimilarity], result of:
              0.0207691 = score(doc=1178,freq=2.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.13412495 = fieldWeight in 1178, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1178)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    In Jorge Luis Borges's curious short story The Library of Babel, the narrator describes an endless collection of books stored from floor to ceiling in a labyrinth of countless hexagonal rooms. The pages of the library's books seem to contain random sequences of letters and spaces; occasionally a few intelligible words emerge in the sea of paper and ink. Nevertheless, readers diligently, and exasperatingly, scan the shelves for coherent passages. The narrator himself has wandered numerous rooms in search of enlightenment, but with resignation he simply awaits his death and burial - which Borges explains (with signature dark humor) consists of being tossed unceremoniously over the library's banister. Borges's nightmare, of course, is a cursed vision of the research methods of disciplines such as literature, history, and philosophy, where the careful reading of books, one after the other, is supposed to lead inexorably to knowledge and understanding. Computer scientists would approach Borges's library far differently. Employing the information theory that forms the basis for search engines and other computerized techniques for assessing in one fell swoop large masses of documents, they would quickly realize the collection's incoherence though sampling and statistical methods - and wisely start looking for the library's exit. These computational methods, which allow us to find patterns, determine relationships, categorize documents, and extract information from massive corpuses, will form the basis for new tools for research in the humanities and other disciplines in the coming decade. For the past three years I have been experimenting with how to provide such end-user tools - that is, tools that harness the power of vast electronic collections while hiding much of their complicated technical plumbing. In particular, I have made extensive use of the application programming interfaces (APIs) the leading search engines provide for programmers to query their databases directly (from server to server without using their web interfaces). In addition, I have explored how one might extract information from large digital collections, from the well-curated lexicographic database WordNet to the democratic (and poorly curated) online reference work Wikipedia. While processing these digital corpuses is currently an imperfect science, even now useful tools can be created by combining various collections and methods for searching and analyzing them. And more importantly, these nascent services suggest a future in which information can be gleaned from, and sense can be made out of, even imperfect digital libraries of enormous scale. A brief examination of two approaches to data mining large digital collections hints at this future, while also providing some lessons about how to get there.
  8. Mimno, D.; Crane, G.; Jones, A.: Hierarchical catalog records : implementing a FRBR catalog (2005) 0.00
    0.002307678 = product of:
      0.006923034 = sum of:
        0.006923034 = product of:
          0.0207691 = sum of:
            0.0207691 = weight(_text_:online in 1183) [ClassicSimilarity], result of:
              0.0207691 = score(doc=1183,freq=2.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.13412495 = fieldWeight in 1183, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1183)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    IFLA's Functional Requirements for Bibliographic Records (FRBR) lay the foundation for a new generation of cataloging systems that recognize the difference between a particular work (e.g., Moby Dick), diverse expressions of that work (e.g., translations into German, Japanese and other languages), different versions of the same basic text (e.g., the Modern Library Classics vs. Penguin editions), and particular items (a copy of Moby Dick on the shelf). Much work has gone into finding ways to infer FRBR relationships between existing catalog records and modifying catalog interfaces to display those relationships. Relatively little work, however, has gone into exploring the creation of catalog records that are inherently based on the FRBR hierarchy of works, expressions, manifestations, and items. The Perseus Digital Library has created a new catalog that implements such a system for a small collection that includes many works with multiple versions. We have used this catalog to explore some of the implications of hierarchical catalog records for searching and browsing. Current online library catalog interfaces present many problems for searching. One commonly cited failure is the inability to find and collocate all versions of a distinct intellectual work that exist in a collection and the inability to take into account known variations in titles and personal names (Yee 2005). The IFLA Functional Requirements for Bibliographic Records (FRBR) attempts to address some of these failings by introducing the concept of multiple interrelated bibliographic entities (IFLA 1998). In particular, relationships between abstract intellectual works and the various published instances of those works are divided into a four-level hierarchy of works (such as the Aeneid), expressions (Robert Fitzgerald's translation of the Aeneid), manifestations (a particular paperback edition of Robert Fitzgerald's translation of the Aeneid), and items (my copy of a particular paperback edition of Robert Fitzgerald's translation of the Aeneid). In this formulation, each level in the hierarchy "inherits" information from the preceding level. Much of the work on FRBRized catalogs so far has focused on organizing existing records that describe individual physical books. Relatively little work has gone into rethinking what information should be in catalog records, or how the records should relate to each other. It is clear, however, that a more "native" FRBR catalog would include separate records for works, expressions, manifestations, and items. In this way, all information about a work would be centralized in one record. Records for subsequent expressions of that work would add only the information specific to each expression: Samuel Butler's translation of the Iliad does not need to repeat the fact that the work was written by Homer. This approach has certain inherent advantages for collections with many versions of the same works: new publications can be cataloged more quickly, and records can be stored and updated more efficiently.
  9. Buckland, M.; Lancaster, L.: Combining place, time, and topic : the Electronic Cultural Atlas Initiative (2004) 0.00
    0.002307678 = product of:
      0.006923034 = sum of:
        0.006923034 = product of:
          0.0207691 = sum of:
            0.0207691 = weight(_text_:online in 1194) [ClassicSimilarity], result of:
              0.0207691 = score(doc=1194,freq=2.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.13412495 = fieldWeight in 1194, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1194)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    The Electronic Cultural Atlas Initiative was formed to encourage scholarly communication and the sharing of data among researchers who emphasize the relationships between place, time, and topic in the study of culture and history. In an effort to develop better tools and practices, The Electronic Cultural Atlas Initiative has sponsored the collaborative development of software for downloading and editing geo-temporal data to create dynamic maps, a clearinghouse of shared datasets accessible through a map-based interface, projects on format and content standards for gazetteers and time period directories, studies to improve geo-temporal aspects in online catalogs, good practice guidelines for preparing e-publications with dynamic geo-temporal displays, and numerous international conferences. The Electronic Cultural Atlas Initiative (ECAI) grew out of discussions among an international group of scholars interested in religious history and area studies. It was established as a unit under the Dean of International and Area Studies at the University of California, Berkeley in 1997. ECAI's mission is to promote an international collaborative effort to transform humanities scholarship through use of the digital environment to share data and by placing greater emphasis on the notions of place and time. Professor Lewis Lancaster is the Director. Professor Michael Buckland, with a library and information studies background, joined the effort as Co-Director in 2000. Assistance from the Lilly Foundation, the California Digital Library (University of California), and other sources has enabled ECAI to nurture a community; to develop a catalog ("clearinghouse") of Internet-accessible georeferenced resources; to support the development of software for obtaining, editing, manipulating, and dynamically visualizing geo-temporally encoded data; and to undertake research and development projects as needs and resources determine. Several hundred scholars worldwide, from a wide range of disciplines, are informally affiliated with ECAI, all interested in shared use of historical and cultural data. The Academia Sinica (Taiwan), The British Library, and the Arts and Humanities Data Service (UK) are among the well-known affiliates. However, ECAI mainly comprises individual scholars and small teams working on their own small projects on a very wide range of cultural, social, and historical topics. Numerous specialist committees have been fostering standardization and collaboration by area and by themes such as trade-routes, cities, religion, and sacred sites.
  10. Kenney, A.R.; McGovern, N.Y.; Martinez, I.T.; Heidig, L.J.: Google meets eBay : what academic librarians can learn from alternative information providers (2003) 0.00
    0.002307678 = product of:
      0.006923034 = sum of:
        0.006923034 = product of:
          0.0207691 = sum of:
            0.0207691 = weight(_text_:online in 1200) [ClassicSimilarity], result of:
              0.0207691 = score(doc=1200,freq=2.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.13412495 = fieldWeight in 1200, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1200)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    In April 2002, the dominant Internet search engine, GoogleT, introduced a beta version of its expert service, Google Answers, with little fanfare. Almost immediately the buzz within the information community focused on implications for reference librarians. Google had already been lauded as the cheaper and faster alternative for finding information, and declining reference statistics and Online Public Access Catalog (OPAC) use in academic libraries had been attributed in part to its popularity. One estimate suggests that the Google search engine handles more questions in a day and a half than all the libraries in the country provide in a year. Indeed, Craig Silverstein, Google's Director of Technology, indicated that the raison d'être for the search engine was to "seem as smart as a reference librarian," even as he acknowledged that this goal was "hundreds of years away". Bill Arms had reached a similar conclusion regarding the more nuanced reference functions in a thought-provoking article in this journal on automating digital libraries. But with the launch of Google Answers, the power of "brute force computing" and simple algorithms could be combined with human intelligence to represent a market-driven alternative to library reference services. Google Answers is part of a much larger trend to provide networked reference assistance. Expert services have sprung up in both the commercial and non-profit sector. Libraries too have responded to the Web, providing a suite of services through the virtual reference desk (VRD) movement, from email reference to chat reference to collaborative services that span the globe. As the Internet's content continues to grow and deepen - encompassing over 40 million web sites - it has been met by a groundswell of services to find and filter information. These services include an extensive range from free to fee-based, cost-recovery to for-profit, and library providers to other information providers - both new and traditional. As academic libraries look towards the future in a dynamic and competitive information landscape, what implications do these services have for their programs, and what can be learned from them to improve library offerings? This paper presents the results of a modest study conducted by Cornell University Library (CUL) to compare and contrast its digital reference services with those of Google Answers. The study provided an opportunity for librarians to shift their focus from fearing the impact of Google, as usurper of the library's role and diluter of the academic experience, to gaining insights into how Google's approach to service development and delivery has made it so attractive.
  11. Hill, L.: New Protocols for Gazetteer and Thesaurus Services (2002) 0.00
    0.002307678 = product of:
      0.006923034 = sum of:
        0.006923034 = product of:
          0.0207691 = sum of:
            0.0207691 = weight(_text_:online in 1206) [ClassicSimilarity], result of:
              0.0207691 = score(doc=1206,freq=2.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.13412495 = fieldWeight in 1206, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1206)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    The Alexandria Digital Library Project announces the online publication of two protocols to support querying and response interactions using distributed services: one for gazetteers and one for thesauri. These protocols have been developed for our own purposes and also to support the general interoperability of gazetteers and thesauri on the web. See <http://www.alexandria.ucsb.edu/~gjanee/gazetteer/> and <http://www.alexandria.ucsb.edu/~gjanee/thesaurus/>. For the gazetteer protocol, we have provided a page of test forms that can be used to experiment with the operational functions of the protocol in accessing two gazetteers: the ADL Gazetteer and the ESRI Gazetteer (ESRI has participated in the development of the gazetteer protocol). We are in the process of developing a thesaurus server and a simple client to demonstrate the use of the thesaurus protocol. We are soliciting comments on both protocols. Please remember that we are seeking protocols that are essentially "simple" and easy to implement and that support basic operations - they should not duplicate all of the functions of specialized gazetteer and thesaurus interfaces. We continue to discuss ways of handling various issues and to further develop the protocols. For the thesaurus protocol, outstanding issues include the treatment of multilingual thesauri and the degree to which the language attribute should be supported; whether the Scope Note element should be changed to a repeatable Note element; the best way to handle the hierarchical report for multi-hierarchies where portions of the hierarchy are repeated; and whether support for searching by term identifiers is redundant and unnecessary given that the terms themselves are unique within a thesaurus. For the gazetteer protocol, we continue to work on validation of query and report XML documents and on implementing the part of the protocol designed to support the submission of new entries to a gazetteer. We would like to encourage open discussion of these protocols through the NKOS discussion list (see the NKOS webpage at <http://nkos.slis.kent.edu/>) and the CGGR-L discussion list that focuses on gazetteer development (see ADL Gazetteer Development page at <http://www.alexandria.ucsb.edu/gazetteer>).
  12. Hill, L.L.; Frew, J.; Zheng, Q.: Geographic names : the implementation of a gazetteer in a georeferenced digital library (1999) 0.00
    0.002307678 = product of:
      0.006923034 = sum of:
        0.006923034 = product of:
          0.0207691 = sum of:
            0.0207691 = weight(_text_:online in 1240) [ClassicSimilarity], result of:
              0.0207691 = score(doc=1240,freq=2.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.13412495 = fieldWeight in 1240, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1240)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    The Alexandria Digital Library (ADL) Project has developed a content standard for gazetteer objects and a hierarchical type scheme for geographic features. Both of these developments are based on ADL experience with an earlier gazetteer component for the Library, based on two gazetteers maintained by the U.S. federal government. We define the minimum components of a gazetteer entry as (1) a geographic name, (2) a geographic location represented by coordinates, and (3) a type designation. With these attributes, a gazetteer can function as a tool for indirect spatial location identification through names and types. The ADL Gazetteer Content Standard supports contribution and sharing of gazetteer entries with rich descriptions beyond the minimum requirements. This paper describes the content standard, the feature type thesaurus, and the implementation and research issues. A gazetteer is list of geographic names, together with their geographic locations and other descriptive information. A geographic name is a proper name for a geographic place and feature, such as Santa Barbara County, Mount Washington, St. Francis Hospital, and Southern California. There are many types of printed gazetteers. For example, the New York Times Atlas has a gazetteer section that can be used to look up a geographic name and find the page(s) and grid reference(s) where the corresponding feature is shown. Some gazetteers provide information about places and features; for example, a history of the locale, population data, physical data such as elevation, or the pronunciation of the name. Some lists of geographic names are available as hierarchical term sets (thesauri) designed for information retreival; these are used to describe bibliographic or museum materials. Examples include the authority files of the U.S. Library of Congress and the GeoRef Thesaurus produced by the American Geological Institute. The Getty Museum has recently made their Thesaurus of Geographic Names available online. This is a major project to develop a controlled vocabulary of current and historical names to describe (i.e., catalog) art and architecture literature. U.S. federal government mapping agencies maintain gazetteers containing the official names of places and/or the names that appear on map series. Examples include the U.S. Geological Survey's Geographic Names Information System (GNIS) and the National Imagery and Mapping Agency's Geographic Names Processing System (GNPS). Both of these are maintained in cooperation with the U.S. Board of Geographic Names (BGN). Many other examples could be cited -- for local areas, for other countries, and for special purposes. There is remarkable diversity in approaches to the description of geographic places and no standardization beyond authoritative sources for the geographic names themselves.
  13. Atkins, H.: ¬The ISI® Web of Science® - links and electronic journals : how links work today in the Web of Science, and the challenges posed by electronic journals (1999) 0.00
    0.002307678 = product of:
      0.006923034 = sum of:
        0.006923034 = product of:
          0.0207691 = sum of:
            0.0207691 = weight(_text_:online in 1246) [ClassicSimilarity], result of:
              0.0207691 = score(doc=1246,freq=2.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.13412495 = fieldWeight in 1246, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1246)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    Since their inception in the early 1960s the strength and unique aspect of the ISI citation indexes has been their ability to illustrate the conceptual relationships between scholarly documents. When authors create reference lists for their papers, they make explicit links between their own, current work and the prior work of others. The exact nature of these links may not be expressed in the references themselves, and the motivation behind them may vary (this has been the subject of much discussion over the years), but the links embodied in references do exist. Over the past 30+ years, technology has allowed ISI to make the presentation of citation searching increasingly accessible to users of our products. Citation searching and link tracking moved from being rather cumbersome in print, to being direct and efficient (albeit non-intuitive) online, to being somewhat more user-friendly in CD format. But it is the confluence of the hypertext link and development of Web browsers that has enabled us to present to users a new form of citation product -- the Web of Science -- that is intuitive and makes citation indexing conceptually accessible. A cited reference search begins with a known, important (or at least relevant) document used as the search term. The search allows one to identify subsequent articles that have cited that document. This feature adds the dimension of prospective searching to the usual retrospective searching that all bibliographic indexes provide. Citation indexing is a prime example of a concept before its time - important enough to be used in the meantime by those sufficiently motivated, but just waiting for the right technology to come along to expand its use. While it was possible to follow citation links in earlier citation index formats, this required a level of effort on the part of users that was often just too much to ask of the casual user. In the citation indexes as presented in the Web of Science, the relationship between citing and cited documents is evident to users, and a click of the mouse is all it takes to follow a citation link. Citation connections are established between the published papers being indexed from the 8,000+ journals ISI covers and the items their reference lists contain during the data capture process. It is the standardized capture of each of the references included with these documents that enables us to provide the citation searching feature in all the citation index formats, as well as both internal and external links in the Web of Science.
  14. Putkey, T.: Using SKOS to express faceted classification on the Semantic Web (2011) 0.00
    0.002307678 = product of:
      0.006923034 = sum of:
        0.006923034 = product of:
          0.0207691 = sum of:
            0.0207691 = weight(_text_:online in 311) [ClassicSimilarity], result of:
              0.0207691 = score(doc=311,freq=2.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.13412495 = fieldWeight in 311, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.03125 = fieldNorm(doc=311)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Issue
    [Online: http://unllib.unl.edu/LPP/].
  15. Lavoie, B.; Connaway, L.S.; Dempsey, L.: Anatomy of aggregate collections : the example of Google print for libraries (2005) 0.00
    0.0023042914 = product of:
      0.0069128736 = sum of:
        0.0069128736 = product of:
          0.02073862 = sum of:
            0.02073862 = weight(_text_:22 in 1184) [ClassicSimilarity], result of:
              0.02073862 = score(doc=1184,freq=2.0), product of:
                0.17867287 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051022716 = queryNorm
                0.116070345 = fieldWeight in 1184, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1184)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Date
    26.12.2011 14:08:22
  16. Bastos Vieira, S.; DeBrito, M.; Mustafa El Hadi, W.; Zumer, M.: Developing imaged KOS with the FRSAD Model : a conceptual methodology (2016) 0.00
    0.0022925164 = product of:
      0.006877549 = sum of:
        0.006877549 = product of:
          0.020632647 = sum of:
            0.020632647 = weight(_text_:retrieval in 3109) [ClassicSimilarity], result of:
              0.020632647 = score(doc=3109,freq=2.0), product of:
                0.15433937 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.051022716 = queryNorm
                0.13368362 = fieldWeight in 3109, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3109)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    This proposal presents the methodology of indexing with images suggested by De Brito and Caribé (2015). The imagetic model is used as a compatible mechanism with FRSAD for a global information share and use of subject data, both within the library sector and beyond. The conceptual model of imagetic indexing shows how images are related to topics and 'key-images' are interpreted as nomens to implement the FRSAD model. Indexing with images consists of using images instead of key-words or descriptors, to represent and organize information. Implementing the imaged navigation in OPACs denotes multiple advantages derived from this rethinking the OPAC anew, since we are looking forward to sharing concepts within the subject authority data. Images, carrying linguistic objects, permeate inter-social and cultural concepts. In practice it includes translated metadata, symmetrical multilingual thesaurus, or any traditional indexing tools. iOPAC embodies efforts focused on conceptual levels as expected from librarians. Imaged interfaces are more intuitive since users do not need specific training for information retrieval, offering easier comprehension of indexing codes, larger conceptual portability of descriptors (as images), and a better interoperability between discourse codes and indexing competences affecting positively social and cultural interoperability. The imagetic methodology deploys R&D fields for more suitable interfaces taking into consideration users with specific needs such as deafness and illiteracy. This methodology arouse questions about the paradigms of the primacy of orality in information systems and pave the way to a legitimacy of multiple perspectives in document indexing by suggesting a more universal communication system based on images. Interdisciplinarity in neurosciences, linguistics and information sciences would be desirable competencies for further investigations about he nature of cognitive processes in information organization and classification while developing assistive KOS for individuals with communication problems, such autism and deafness.
  17. Kubiszewski, I.; Cleveland, C.J.: ¬The Encyclopedia of Earth (2007) 0.00
    0.0020192184 = product of:
      0.006057655 = sum of:
        0.006057655 = product of:
          0.018172964 = sum of:
            0.018172964 = weight(_text_:online in 1170) [ClassicSimilarity], result of:
              0.018172964 = score(doc=1170,freq=2.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.11735933 = fieldWeight in 1170, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1170)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    This illustrates a stark reality of the Web. There are many resources for environmental content, but there is no central repository of authoritative information that meets the needs of diverse user communities. The Encyclopedia of Earth aims to fill that niche by providing content that is both free and reliable. Still in its infancy, the EoE already is an integral part of the emerging effort to increase free and open access to trusted information on the Web. It is a trusted content source for authoritative indexes such as the Online Access to Research in the Environment Initiative, the Health InterNetwork Access to Research Initiative, the Open Education Resources Commons, Scirus, DLESE, WiserEarth, among others. Our initial Content Partners include the American Institute of Physics, the University of California Museum of Paleontology, TeacherServe®, the U.S. Geological Survey, the International Arctic Science Committee, the World Wildlife Fund, Conservation International, the Biodiversity Institute of Ontario, and the United Nations Environment Programme, to name just a few. The full partner list here can be found at <http://www.eoearth.org/article/Content_Partners>. We have a diversity of article types including standard subject articles, biographies, place-based entries, country profiles, and environmental classics. We recently launched our E-Book series, full-text, fully searchable books with internal hyperlinks to EoE articles. The eBooks include new releases by distinguished scholars as well as classics such as Walden and On the Origin of Species. Because history can be an important guide to the future, we have added an Environmental Classics section that includes such historical works as Energy from Fossil Fuels by M. King Hubbert and Undersea by Rachel Carson. Our services and features will soon be expanded. The EoE will soon be available in different languages giving a wider range of users access, users will be able to search it geographically or by a well-defined, expert created taxonomy, and teachers will be able to use the EoE to create unique curriculum for their courses.
  18. Lenzen, M.: Vor der Quadratwurzel steht die Quadratzahl (2015) 0.00
    0.0020192184 = product of:
      0.006057655 = sum of:
        0.006057655 = product of:
          0.018172964 = sum of:
            0.018172964 = weight(_text_:online in 1647) [ClassicSimilarity], result of:
              0.018172964 = score(doc=1647,freq=2.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.11735933 = fieldWeight in 1647, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1647)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Source
    http://www.faz.net/aktuell/feuilleton/forschung-und-lehre/uni-mainz-stellt-publikationen-von-hirnforschern-online-13379697.html
  19. Bertolucci, K.: Happiness is taxonomy : four structures for Snoopy - libraries' method of categorizing and classification (2003) 0.00
    0.0017307586 = product of:
      0.0051922756 = sum of:
        0.0051922756 = product of:
          0.015576826 = sum of:
            0.015576826 = weight(_text_:online in 1212) [ClassicSimilarity], result of:
              0.015576826 = score(doc=1212,freq=2.0), product of:
                0.1548489 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.051022716 = queryNorm
                0.100593716 = fieldWeight in 1212, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1212)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    Dewey and the Library of Congress The late 19th and early 20th centuries were a hotbed of intellectual activity for library categorizers. First Melvil Dewey developed his decimal system. Then the Library of Congress (LC) adapted Charles Ammi Cutter's alphanumeric system for its collection. Dewey, the only librarian popularly known for librarianship, had a healthy ego and placed information science at the very beginning of his classifications. The librarians at LC followed Cutter and relegated their profession to the back of their own bus, in the Zs. These two systems became the primary classifications accepted by the library community. I was once chastised at an SLA meeting for daring to design my own systems, and library schools that mainly train people for public and academic institutions reinforce this idea. In addition, LC provides cataloging and call numbers for almost every book commercially published in the United States and quite a few international publications. This is a seductive strategy for libraries that have little money and little time. These two systems contain drawbacks for special libraries. Let's see how they treat Snoopy. I'll be using Dewey for this exercise. Dewey has an index, which facilitates classification analysis. In addition, LC is a larger system, and we have space considerations here. However, other than length, call number building, and self-esteem, there is not much difference in the two theories. Figure 2 shows selected Dewey classifications for Snoopy, beagles, dogs, and animals (Melvil Dewey. Dewey Decimal Classification and Relative Index. 21st ed. Edited by Joan S. Mitchell, et al. Albany, NY: OCLC Online Computer Library Center, 1996). The call numbers are removed to emphasize hierarchy rather than notation. There are 234 categories. Both Dewey and LC are designed to describe the whole of human knowledge. For historic reasons, they do this from the perspective of an educated white male in 19th century America. This perspective presents some problems if your specialty is Snoopy. In "Generalities," newspaper cartoon strips are filed away under "Miscellaneous information, advice, amusement." However, a collection of Charles Schulz cartoons would be shelved way over in "The Arts [right arrow] Drawing and decorative arts," thereby separating two almost equal subjects by a very wide distance. The generic vocabulary required to describe all of human knowledge is also problematic for specialists. In "The Arts [right arrow] Standard subdivisions of fine and decorative arts and iconography," there are five synonyms for miscellaneous before we get to a real subject. Then it's another six facets to get to the dogs.

Years

Languages

  • d 297
  • e 137
  • a 1
  • el 1
  • es 1
  • i 1
  • More… Less…