Search (22530 results, page 1127 of 1127)

  1. Progress in visual information access and retrieval (1999) 0.00
    1.08939676E-4 = product of:
      0.0026145522 = sum of:
        0.0026145522 = product of:
          0.007843656 = sum of:
            0.007843656 = weight(_text_:p in 839) [ClassicSimilarity], result of:
              0.007843656 = score(doc=839,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.099312946 = fieldWeight in 839, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=839)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
    Content
    Enthält die Beiträge: Part I-Foundations of Access to Visual Information Intellectual Access to Images (Hsin-liang Chen and Edie M. Rasmussen) - Image Retrieval as Linguistic and Nonlinguistic Visual Model Matching (P. Bryan Heidorn) - Computer Vision Tools for Finding Images and Video Sequences (D. A. Forsyth) Part II-Implementation and Evaluation Securing Digital Image Assets in Museums and Libraries: A Risk Management Approach (Teresa Grose Beamsley) - Getting the Picture: Observations from the Library of Congress on Providing Access to Pictorial Images (Caroline R. Arms) - Recent Developments in Cultural Heritage Image Databases: Directions for User-Centered Design (Christie Stephenson) - Evaluation of Image Retrieval Systems: Role of User Feedback (Samantha K. Hastings) Part III-Experimental Approaches Information Retrieval Beyond the Text Document (Yong Rui, Michael Ortega, Thomas S. Huang, and Sharad Mehrotra) - Precise and Efficient Retrieval of Captioned Images: The MARIE Project (Neil C. Rowe) - Exploiting Multimodal Context in Image Retrieval (Rohini K. Srihari and Zhongfei Zhang)
  2. Smiraglia, R.P.: Curating and virtual shelves : an editorial (2006) 0.00
    1.08939676E-4 = product of:
      0.0026145522 = sum of:
        0.0026145522 = product of:
          0.007843656 = sum of:
            0.007843656 = weight(_text_:p in 409) [ClassicSimilarity], result of:
              0.007843656 = score(doc=409,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.099312946 = fieldWeight in 409, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=409)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
    Content
    Librarianship incorporates the tools of knowledge organization as part of its role as cultural disseminator. Subject headings and classification were both intended by their 19`h century promulgators - perhaps most notably Dewey and Cutter - to facilitate learning by grouping materials of high quality together. We might call this enhanced serendipity if we think it happens by accident or act of fate, or we might call it curatorship if we realize the responsibility inherent in our social role. The cataloger's job always has been to place each work sensitively among other works related to it, and to make the relationships explicit to facilitate and even encourage selection (see Miksa 1983). Schallier (2004) reported on the use of classification in an online catalog to enhance just such a curatorial purpose. UDC classification codes were exploded into linguistic strings to allow users to search, not just for a given term, but for the terms that occur around it - that is, terms that are adjacent in the classification. These displays are used alongside LCSH to provide enhanced-serendipity for users. What caught my attention was the intention of the project (p. 271): UDC permits librarians to build virtual library shelves, where a document's subjects can be described in thematic categories rather than in detailed verbal terms. And: It is our experience that most end users are not familiar with large controlled vocabularies. UDC could be an answer to this, since its alphanumeric makeup could be used to build a tree structure of terms, which would guide end users in their searchers. There are other implications from this project, including background linkage from UDC codes that drive the "virtual shelves" to subject terms that drive the initial classification. Knowledge organization has consequences in both theory and application."
  3. Warner, J.: Humanizing information technology (2004) 0.00
    1.08939676E-4 = product of:
      0.0026145522 = sum of:
        0.0026145522 = product of:
          0.007843656 = sum of:
            0.007843656 = weight(_text_:p in 438) [ClassicSimilarity], result of:
              0.007843656 = score(doc=438,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.099312946 = fieldWeight in 438, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=438)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
    Footnote
    Like Daniel Bell, the author of The Coming of Post-Industrial Society (1973), who used aspects of Marx's thinking as the basis for his social forecasting models, Warner uses Marxist thought as a tool for social and historical analysis. Unlike Bell, Warner's approach to Marx tends to be doctrinaire. As a result, "An Information View of History" and "Origins of the Human Brain," two of the essays in which Warner sets out to establish the connections between information science and information technology, are less successful. Warner argues, "the classic source for an understanding of technology as a human construction is Marx," and that "a Marxian perspective an information technology could be of high marginal Utility," noting additionally that with the exception of Norbert Wiener and John Desmond Bernal, "there has only been a limited penetration of Marxism into information science" (p. 9). But Warner's efforts to persuade the reader that these views are cogent never go beyond academic protocol. Nor does his support for the assertion that the second half of the 19th century was the critical period for innovation and diffusion of modern information technologies. The closing essay, "Whither Information Science?" is particularly disappointing, in part, because the preface and opening chapters of the book promised more than was delivered at the end. Warner asserts that the theoretical framework supporting information science is negligible, and that the discipline is limited even further by the fact that many of its members do not recognize or understand the effects of such a limitation. However cogent the charges may be, none of this is news. But the essay fails most notably because Warner does not have any new directions to offer, save that information scientists should pay closer artention to what is going an in allied disciplines. Moreover, he does not seem to understand that at its heart the "information revolution" is not about the machines, but about the growing legions of men and women who can and do write programming code to exert control over and find new uses for these devices. Nor does he seem to understand that information science, in the grip of what he terms a "quasi-global crisis," suffers grievously because it is a community situated not at the center but rather an the periphery of this revolution."
  4. Crane, G.; Jones, A.: Text, information, knowledge and the evolving record of humanity (2006) 0.00
    1.08939676E-4 = product of:
      0.0026145522 = sum of:
        0.0026145522 = product of:
          0.007843656 = sum of:
            0.007843656 = weight(_text_:p in 1182) [ClassicSimilarity], result of:
              0.007843656 = score(doc=1182,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.099312946 = fieldWeight in 1182, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=1182)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
    Abstract
    Consider a sentence such as "the current price of tea in China is 35 cents per pound." In a library with millions of books we might find many statements of the above form that we could capture today with relatively simple rules: rather than pursuing every variation of a statement, programs can wait, like predators at a water hole, for their informational prey to reappear in a standard linguistic pattern. We can make inferences from sentences such as "NAME1 born at NAME2 in DATE" that NAME more likely than not represents a person and NAME a place and then convert the statement into a proposition about a person born at a given place and time. The changing price of tea in China, pedestrian birth and death dates, or other basic statements may not be truth and beauty in the Phaedrus, but a digital library that could plot the prices of various commodities in different markets over time, plot the various lifetimes of individuals, or extract and classify many events would be very useful. Services such as the Syllabus Finder1 and H-Bot2 (which Dan Cohen describes elsewhere in this issue of D-Lib) represent examples of information extraction already in use. H-Bot, in particular, builds on our evolving ability to extract information from very large corpora such as the billions of web pages available through the Google API. Aside from identifying higher order statements, however, users also want to search and browse named entities: they want to read about "C. P. E. Bach" rather than his father "Johann Sebastian" or about "Cambridge, Maryland", without hearing about "Cambridge, Massachusetts", Cambridge in the UK or any of the other Cambridges scattered around the world. Named entity identification is a well-established area with an ongoing literature. The Natural Language Processing Research Group at the University of Sheffield has developed its open source Generalized Architecture for Text Engineering (GATE) for years, while IBM's Unstructured Information Analysis and Search (UIMA) is "available as open source software to provide a common foundation for industry and academia." Powerful tools are thus freely available and more demanding users can draw upon published literature to develop their own systems. Major search engines such as Google and Yahoo also integrate increasingly sophisticated tools to categorize and identify places. The software resources are rich and expanding. The reference works on which these systems depend, however, are ill-suited for historical analysis. First, simple gazetteers and similar authority lists quickly grow too big for useful information extraction. They provide us with potential entities against which to match textual references, but existing electronic reference works assume that human readers can use their knowledge of geography and of the immediate context to pick the right Boston from the Bostons in the Getty Thesaurus of Geographic Names (TGN), but, with the crucial exception of geographic location, the TGN records do not provide any machine readable clues: we cannot tell which Bostons are large or small. If we are analyzing a document published in 1818, we cannot filter out those places that did not yet exist or that had different names: "Jefferson Davis" is not the name of a parish in Louisiana (tgn,2000880) or a county in Mississippi (tgn,2001118) until after the Civil War.
  5. Kochtanek, T.R.; Matthews, J.R.: Library information systems : from library automation to distributed information systems (2002) 0.00
    1.08939676E-4 = product of:
      0.0026145522 = sum of:
        0.0026145522 = product of:
          0.007843656 = sum of:
            0.007843656 = weight(_text_:p in 1792) [ClassicSimilarity], result of:
              0.007843656 = score(doc=1792,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.099312946 = fieldWeight in 1792, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=1792)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
    Footnote
    Rez. in: JASIST 54(2003) no.12, S.1166-1167 (Brenda Chawner): "Kochtanek and Matthews have written a welcome addition to the small set of introductory texts an applications of information technology to library and information Services. The book has fourteen chapters grouped into four sections: "The Broader Context," "The Technologies," "Management Issues," and "Future Considerations." Two chapters provide the broad content, with the first giving a historical overview of the development and adoption of "library information systems." Kochtanek and Matthews define this as "a wide array of solutions that previously might have been considered separate industries with distinctly different marketplaces" (p. 3), referring specifically to integrated library systems (ILS, and offen called library management systems in this part of the world), and online databases, plus the more recent developments of Web-based resources, digital libraries, ebooks, and ejournals. They characterize technology adoption patterns in libraries as ranging from "bleeding edge" to "leading edge" to "in the wedge" to "trailing edge"-this is a catchy restatement of adopter categories from Rogers' diffusion of innovation theory, where they are more conventionally known as "early adopters," "early majority," "late majority," and "laggards." This chapter concludes with a look at more general technology trends that have affected library applications, including developments in hardware (moving from mainframes to minicomputers to personal Computers), changes in software development (from in-house to packages), and developments in communications technology (from dedicated host Computers to more open networks to the current distributed environment found with the Internet). This is followed by a chapter describing the ILS and online database industries in some detail. "The Technologies" begins with a chapter an the structure and functionality of integrated library systems, which also includes a brief discussion of precision versus recall, managing access to internal documents, indexing and searching, and catalogue maintenance. This is followed by a chapter an open systems, which concludes with a useful list of questions to consider to determine an organization's readiness to adopt open source solutions. As one world expect, this section also includes a detailed chapter an telecommunications and networking, which includes types of networks, transmission media, network topologies, switching techniques (ranging from dial up and leased lines to ISDN/DSL, frame relay, and ATM). It concludes with a chapter an the role and importance of standards, which covers the need for standards and standards organizations, and gives examples of different types of standards, such as MARC, Dublin Core, Z39.50, and markup standards such as SGML, HTML, and XML. Unicode is also covered but only briefly. This section world be strengthened by a chapter an hardware concepts-the authors assume that their reader is already familiar with these, which may not be true in all cases (for example, the phrase "client-Server" is first used an page 11, but only given a brief definition in the glossary). Burke's Library Technology Companion: A Basic Guide for Library Staff (New York: Neal-Schuman, 2001) might be useful to fill this gap at an introductory level, and Saffady's Introduction to Automation for Librarians, 4th ed. (Chicago: American Library Association, 1999) world be better for those interested in more detail. The final two sections, however, are the book's real strength, with a strong focus an management issues, and this content distinguishes it from other books an this topic such as Ferguson and Hebels Computers for Librarians: an Introduction to Systems and Applications (Waggawagga, NSW: Centre for Information Studies, Charles Sturt University, 1998). ...
  6. Culture and identity in knowledge organization : Proceedings of the Tenth International ISKO Conference 5-8 August 2008, Montreal, Canada (2008) 0.00
    1.08939676E-4 = product of:
      0.0026145522 = sum of:
        0.0026145522 = product of:
          0.007843656 = sum of:
            0.007843656 = weight(_text_:p in 2494) [ClassicSimilarity], result of:
              0.007843656 = score(doc=2494,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.099312946 = fieldWeight in 2494, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=2494)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
    Content
    EPISTEMOLOGICAL FOUNDATIONS OF KNOWLEDGE ORGANIZATION H. Peter Ohly. Knowledge Organization Pro and Retrospective. Judith Simon. Knowledge and Trust in Epistemology and Social Software/Knowledge Technologies. - D. Grant Campbell. Derrida, Logocentrism, and the Concept of Warrant on the Semantic Web. - Jian Qin. Controlled Semantics Versus Social Semantics: An Epistemological Analysis. - Hope A. Olson. Wind and Rain and Dark of Night: Classification in Scientific Discourse Communities. - Thomas M. Dousa. Empirical Observation, Rational Structures, and Pragmatist Aims: Epistemology and Method in Julius Otto Kaiser's Theory of Systematic Indexing. - Richard P. Smiraglia. Noesis: Perception and Every Day Classification. Birger Hjorland. Deliberate Bias in Knowledge Organization? Joseph T. Tennis and Elin K. Jacob. Toward a Theory of Structure in Information Organization Frameworks. - Jack Andersen. Knowledge Organization as a Cultural Form: From Knowledge Organization to Knowledge Design. - Hur-Li Lee. Origins of the Main Classes in the First Chinese Bibliographie Classification. NON-TEXTUAL MATERIALS Abby Goodrum, Ellen Hibbard, Deborah Fels and Kathryn Woodcock. The Creation of Keysigns American Sign Language Metadata. - Ulrika Kjellman. Visual Knowledge Organization: Towards an International Standard or a Local Institutional Practice?
  7. Next generation search engines : advanced models for information retrieval (2012) 0.00
    1.08939676E-4 = product of:
      0.0026145522 = sum of:
        0.0026145522 = product of:
          0.007843656 = sum of:
            0.007843656 = weight(_text_:p in 357) [ClassicSimilarity], result of:
              0.007843656 = score(doc=357,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.099312946 = fieldWeight in 357, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=357)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
    Content
    Enthält die Beiträge: Das, A., A. Jain: Indexing the World Wide Web: the journey so far. Ke, W.: Decentralized search and the clustering paradox in large scale information networks. Roux, M.: Metadata for search engines: what can be learned from e-Sciences? Fluhr, C.: Crosslingual access to photo databases. Djioua, B., J.-P. Desclés u. M. Alrahabi: Searching and mining with semantic categories. Ghorbel, H., A. Bahri u. R. Bouaziz: Fuzzy ontologies building platform for Semantic Web: FOB platform. Lassalle, E., E. Lassalle: Semantic models in information retrieval. Berry, M.W., R. Esau u. B. Kiefer: The use of text mining techniques in electronic discovery for legal matters. Sleem-Amer, M., I. Bigorgne u. S. Brizard u.a.: Intelligent semantic search engines for opinion and sentiment mining. Hoeber, O.: Human-centred Web search.
  8. Exploring artificial intelligence in the new millennium (2003) 0.00
    8.715174E-5 = product of:
      0.0020916418 = sum of:
        0.0020916418 = product of:
          0.006274925 = sum of:
            0.006274925 = weight(_text_:p in 2099) [ClassicSimilarity], result of:
              0.006274925 = score(doc=2099,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.079450354 = fieldWeight in 2099, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.015625 = fieldNorm(doc=2099)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
    Footnote
    The book does achieve its aim of being a starting point for someone interested in the state of some areas of AI research at the beginning of the new millennium. The book's most irritating feature is the different writing styles of the authors. The book is organized as a collection of papers similar to a typical graduate survey course packet, and as a result the book does not possess a narrative flow. Also the book contains a number of other major weaknesses such as a lack of an introductory or concluding chapter. The book could greatly benefit from an introductory chapter that would introduce readers to the areas of AI, explain why such a book is needed, and explain why each author's research is important. The manner in which the book currently handles these issues is a preface that talks about some of the above issues in a superficial manner. Also such an introductory chapter could be used to expound an what level of AI mathematical and statistical knowledge is expected from readers in order to gain maximum benefit from this book. A concluding chapter would be useful to readers interested in the other areas of AI not covered by the book, as well as open issues common to all of the research presented. In addition, most of the contributors come exclusively from the computer science field, which heavily slants the work toward the computer science community. A great deal of the research presented is being used by a number of research communities outside of computer science, such as biotechnology and information technology. A wider audience for this book could have been achieved by including a more diverse range of authors showing the interdisciplinary nature of many of these fields. Also the book's editors state, "The reader is expected to have basic knowledge of AI at the level of an introductory course to the field" (p vii), which is not the case for this book. Readers need at least a strong familiarity with many of the core concepts within AI, because a number of the chapters are shallow and terse in their historical overviews. Overall, this book would be a useful tool for a professor putting together a survey course an AI research. Most importantly the book would be useful for eager graduate students in need of a starting point for their research for their thesis. This book is best suited as a reference guide to be used by individuals with a strong familiarity with AI."
  9. Intner, S.S.; Lazinger, S.S.; Weihs, J.: Metadata and its impact on libraries (2005) 0.00
    8.715174E-5 = product of:
      0.0020916418 = sum of:
        0.0020916418 = product of:
          0.006274925 = sum of:
            0.006274925 = weight(_text_:p in 339) [ClassicSimilarity], result of:
              0.006274925 = score(doc=339,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.079450354 = fieldWeight in 339, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.015625 = fieldNorm(doc=339)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    
    Footnote
    Chapter 8 discusses issues of archiving and preserving digital materials. The chapter reiterates, "What is the point of all of this if the resources identified and catalogued are not preserved?" (Gorman, 2003, p. 16). Discussion about preservation and related issues is organized in five sections that successively ask why, what, who, how, and how much of the plethora of digital materials should be archived and preserved. These are not easy questions because of media instability and technological obsolescence. Stakeholders in communities with diverse interests compete in terms of which community or representative of a community has an authoritative say in what and how much get archived and preserved. In discussing the above-mentioned questions, the authors once again provide valuable information and lessons from a number of initiatives in Europe, Australia, and from other global initiatives. The Draft Charter on the Preservation of the Digital Heritage and the Guidelines for the Preservation of Digital Heritage, both published by UNESCO, are discussed and some of the preservation principles from the Guidelines are listed. The existing diversity in administrative arrangements for these new projects and resources notwithstanding, the impact on content produced for online reserves through work done in digital projects and from the use of metadata and the impact on levels of reference services and the ensuing need for different models to train users and staff is undeniable. In terms of education and training, formal coursework, continuing education, and informal and on-the-job training are just some of the available options. The intensity in resources required for cataloguing digital materials, the questions over the quality of digital resources, and the threat of the new digital environment to the survival of the traditional library are all issues quoted by critics and others, however, who are concerned about a balance for planning and resources allocated for traditional or print-based resources and newer digital resources. A number of questions are asked as part of the book's conclusions in Chapter 10. Of these questions, one that touches on all of the rest and upon much of the book's content is the question: What does the future hold for metadata in libraries? Metadata standards are alive and well in many communities of practice, as Chapters 2-6 have demonstrated. The usefulness of metadata continues to be high and innovation in various elements should keep information professionals engaged for decades to come. There is no doubt that metadata have had a tremendous impact in how we organize information for access and in terms of who, how, when, and where contact is made with library services and collections online. Planning and commitment to a diversity of metadata to serve the plethora of needs in communities of practice are paramount for the continued success of many digital projects and for online preservation of our digital heritage."
  10. Lambe, P.: Organising knowledge : taxonomies, knowledge and organisational effectiveness (2007) 0.00
    8.715174E-5 = product of:
      0.0020916418 = sum of:
        0.0020916418 = product of:
          0.006274925 = sum of:
            0.006274925 = weight(_text_:p in 1804) [ClassicSimilarity], result of:
              0.006274925 = score(doc=1804,freq=2.0), product of:
                0.078979194 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.021966046 = queryNorm
                0.079450354 = fieldWeight in 1804, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.015625 = fieldNorm(doc=1804)
          0.33333334 = coord(1/3)
      0.041666668 = coord(1/24)
    

Authors

Languages

Types

Themes

Subjects

Classifications