Search (55 results, page 1 of 3)

  • × theme_ss:"Literaturübersicht"
  1. Thelwall, M.; Vaughan, L.; Björneborn, L.: Webometrics (2004) 0.02
    0.018483635 = product of:
      0.12938544 = sum of:
        0.05661395 = weight(_text_:web in 4279) [ClassicSimilarity], result of:
          0.05661395 = score(doc=4279,freq=18.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.5408555 = fieldWeight in 4279, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4279)
        0.07277149 = weight(_text_:log in 4279) [ClassicSimilarity], result of:
          0.07277149 = score(doc=4279,freq=2.0), product of:
            0.205552 = queryWeight, product of:
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.0320743 = queryNorm
            0.3540296 = fieldWeight in 4279, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4279)
      0.14285715 = coord(2/14)
    
    Abstract
    Webometrics, the quantitative study of Web-related phenomena, emerged from the realization that methods originally designed for bibliometric analysis of scientific journal article citation patterns could be applied to the Web, with commercial search engines providing the raw data. Almind and Ingwersen (1997) defined the field and gave it its name. Other pioneers included Rodriguez Gairin (1997) and Aguillo (1998). Larson (1996) undertook exploratory link structure analysis, as did Rousseau (1997). Webometrics encompasses research from fields beyond information science such as communication studies, statistical physics, and computer science. In this review we concentrate on link analysis, but also cover other aspects of webometrics, including Web log fle analysis. One theme that runs through this chapter is the messiness of Web data and the need for data cleansing heuristics. The uncontrolled Web creates numerous problems in the interpretation of results, for instance, from the automatic creation or replication of links. The loose connection between top-level domain specifications (e.g., com, edu, and org) and their actual content is also a frustrating problem. For example, many .com sites contain noncommercial content, although com is ostensibly the main commercial top-level domain. Indeed, a skeptical researcher could claim that obstacles of this kind are so great that all Web analyses lack value. As will be seen, one response to this view, a view shared by critics of evaluative bibliometrics, is to demonstrate that Web data correlate significantly with some non-Web data in order to prove that the Web data are not wholly random. A practical response has been to develop increasingly sophisticated data cleansing techniques and multiple data analysis methods.
  2. Bar-Ilan, J.: ¬The use of Web search engines in information science research (2003) 0.02
    0.018194899 = product of:
      0.1273643 = sum of:
        0.052257393 = weight(_text_:source in 4271) [ClassicSimilarity], result of:
          0.052257393 = score(doc=4271,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.32864225 = fieldWeight in 4271, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.046875 = fieldNorm(doc=4271)
        0.0751069 = weight(_text_:web in 4271) [ClassicSimilarity], result of:
          0.0751069 = score(doc=4271,freq=22.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.717526 = fieldWeight in 4271, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4271)
      0.14285715 = coord(2/14)
    
    Abstract
    The World Wide Web was created in 1989, but it has already become a major information channel and source, influencing our everyday lives, commercial transactions, and scientific communication, to mention just a few areas. The seventeenth-century philosopher Descartes proclaimed, "I think, therefore I am" (cogito, ergo sum). Today the Web is such an integral part of our lives that we could rephrase Descartes' statement as "I have a Web presence, therefore I am." Because many people, companies, and organizations take this notion seriously, in addition to more substantial reasons for publishing information an the Web, the number of Web pages is in the billions and growing constantly. However, it is not sufficient to have a Web presence; tools that enable users to locate Web pages are needed as well. The major tools for discovering and locating information an the Web are search engines. This review discusses the use of Web search engines in information science research. Before going into detail, we should define the terms "information science," "Web search engine," and "use" in the context of this review.
  3. Denton, W.: Putting facets on the Web : an annotated bibliography (2003) 0.02
    0.0155198 = product of:
      0.07242573 = sum of:
        0.017965734 = weight(_text_:open in 2467) [ClassicSimilarity], result of:
          0.017965734 = score(doc=2467,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.12438454 = fieldWeight in 2467, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.01953125 = fieldNorm(doc=2467)
        0.021773916 = weight(_text_:source in 2467) [ClassicSimilarity], result of:
          0.021773916 = score(doc=2467,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.13693428 = fieldWeight in 2467, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.01953125 = fieldNorm(doc=2467)
        0.03268608 = weight(_text_:web in 2467) [ClassicSimilarity], result of:
          0.03268608 = score(doc=2467,freq=24.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.3122631 = fieldWeight in 2467, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.01953125 = fieldNorm(doc=2467)
      0.21428572 = coord(3/14)
    
    Abstract
    This is a classified, annotated bibliography about how to design faceted classification systems and make them usable on the World Wide Web. It is the first of three works I will be doing. The second, based on the material here and elsewhere, will discuss how to actually make the faceted system and put it online. The third will be a report of how I did just that, what worked, what didn't, and what I learned. Almost every article or book listed here begins with an explanation of what a faceted classification system is, so I won't (but see Steckel in Background below if you don't already know). They all agree that faceted systems are very appropriate for the web. Even pre-web articles (such as Duncan's in Background, below) assert that hypertext and facets will go together well. Combined, it is possible to take a set of documents and classify them or apply subject headings to describe what they are about, then build a navigational structure so that any user, no matter how he or she approaches the material, no matter what his or her goals, can move and search in a way that makes sense to them, but still get to the same useful results as someone else following a different path to the same goal. There is no one way that everyone will always use when looking for information. The more flexible the organization of the information, the more accommodating it is. Facets are more flexible for hypertext browsing than any enumerative or hierarchical system.
    Consider movie listings in newspapers. Most Canadian newspapers list movie showtimes in two large blocks, for the two major theatre chains. The listings are ordered by region (in large cities), then theatre, then movie, and finally by showtime. Anyone wondering where and when a particular movie is playing must scan the complete listings. Determining what movies are playing in the next half hour is very difficult. When movie listings went onto the web, most sites used a simple faceted organization, always with movie name and theatre, and perhaps with region or neighbourhood (thankfully, theatre chains were left out). They make it easy to pick a theatre and see what movies are playing there, or to pick a movie and see what theatres are showing it. To complete the system, the sites should allow users to browse by neighbourhood and showtime, and to order the results in any way they desired. Thus could people easily find answers to such questions as, "Where is the new James Bond movie playing?" "What's showing at the Roxy tonight?" "I'm going to be out in in Little Finland this afternoon with three hours to kill starting at 2 ... is anything interesting playing?" A hypertext, faceted classification system makes more useful information more easily available to the user. Reading the books and articles below in chronological order will show a certain progression: suggestions that faceting and hypertext might work well, confidence that facets would work well if only someone would make such a system, and finally the beginning of serious work on actually designing, building, and testing faceted web sites. There is a solid basis of how to make faceted classifications (see Vickery in Recommended), but their application online is just starting. Work on XFML (see Van Dijck's work in Recommended) the Exchangeable Faceted Metadata Language, will make this easier. If it follows previous patterns, parts of the Internet community will embrace the idea and make open source software available for others to reuse. It will be particularly beneficial if professionals in both information studies and computer science can work together to build working systems, standards, and code. Each can benefit from the other's expertise in what can be a very complicated and technical area. One particularly nice thing about this area of research is that people interested in combining facets and the web often have web sites where they post their writings.
    This bibliography is not meant to be exhaustive, but unfortunately it is not as complete as I wanted. Some books and articles are not be included, but they may be used in my future work. (These include two books and one article by B.C. Vickery: Faceted Classification Schemes (New Brunswick, NJ: Rutgers, 1966), Classification and Indexing in Science, 3rd ed. (London: Butterworths, 1975), and "Knowledge Representation: A Brief Review" (Journal of Documentation 42 no. 3 (September 1986): 145-159; and A.C. Foskett's "The Future of Faceted Classification" in The Future of Classification, edited by Rita Marcella and Arthur Maltby (Aldershot, England: Gower, 2000): 69-80). Nevertheless, I hope this bibliography will be useful for those both new to or familiar with faceted hypertext systems. Some very basic resources are listed, as well as some very advanced ones. Some example web sites are mentioned, but there is no detailed technical discussion of any software. The user interface to any web site is extremely important, and this is briefly mentioned in two or three places (for example the discussion of lawforwa.org (see Example Web Sites)). The larger question of how to display information graphically and with hypertext is outside the scope of this bibliography. There are five sections: Recommended, Background, Not Relevant, Example Web Sites, and Mailing Lists. Background material is either introductory, advanced, or of peripheral interest, and can be read after the Recommended resources if the reader wants to know more. The Not Relevant category contains articles that may appear in bibliographies but are not relevant for my purposes.
  4. Deokattey, S.; Sharma, S.B.K.; Kumar, G.R.; Bhanumurthy, K.: Knowledge organization research : an overview (2015) 0.01
    0.010882379 = product of:
      0.07617664 = sum of:
        0.060966957 = weight(_text_:source in 2092) [ClassicSimilarity], result of:
          0.060966957 = score(doc=2092,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.38341597 = fieldWeight in 2092, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2092)
        0.015209687 = product of:
          0.030419374 = sum of:
            0.030419374 = weight(_text_:22 in 2092) [ClassicSimilarity], result of:
              0.030419374 = score(doc=2092,freq=2.0), product of:
                0.11231873 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0320743 = queryNorm
                0.2708308 = fieldWeight in 2092, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2092)
          0.5 = coord(1/2)
      0.14285715 = coord(2/14)
    
    Abstract
    The object of this literature review is to provide a historical perspective of R and D work in the area of Knowledge Organization (KO). This overview/summarization will provide information on major areas of KO. Journal articles published in core areas of KO: (Classification, Indexing, Thesauri and Taxonomies, Internet and Subject approach to information in the electronic era and Ontologies will be predominantly covered in this literature review. Coverage in this overview may not be completely exhaustive, but it succinctly showcases major developments in the area of KO. This review is a good source of additional reading material on KO apart from prescribed reading material on KO
    Date
    22. 6.2015 16:13:38
  5. Genereux, C.: Building connections : a review of the serials literature 2004 through 2005 (2007) 0.01
    0.00802209 = product of:
      0.05615463 = sum of:
        0.043117758 = weight(_text_:open in 2548) [ClassicSimilarity], result of:
          0.043117758 = score(doc=2548,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.2985229 = fieldWeight in 2548, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.046875 = fieldNorm(doc=2548)
        0.013036874 = product of:
          0.026073748 = sum of:
            0.026073748 = weight(_text_:22 in 2548) [ClassicSimilarity], result of:
              0.026073748 = score(doc=2548,freq=2.0), product of:
                0.11231873 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0320743 = queryNorm
                0.23214069 = fieldWeight in 2548, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2548)
          0.5 = coord(1/2)
      0.14285715 = coord(2/14)
    
    Abstract
    This review of 2004 and 2005 serials literature covers the themes of cost, management, and access. Interwoven through the serials literature of these two years are the importance of collaboration, communication, and linkages between scholars, publishers, subscription agents and other intermediaries, and librarians. The emphasis in the literature is on electronic serials and their impact on publishing, libraries, and vendors. In response to the crisis of escalating journal prices and libraries' dissatisfaction with the Big Deal licensing agreements, Open Access journals and publishing models were promoted. Libraries subscribed to or licensed increasing numbers of electronic serials. As a result, libraries sought ways to better manage licensing and subscription data (not handled by traditional integrated library systems) by implementing electronic resources management systems. In order to provide users with better, faster, and more current information on and access to electronic serials, libraries implemented tools and services to provide A-Z title lists, title by title coverage data, MARC records, and OpenURL link resolvers.
    Date
    10. 9.2000 17:38:22
  6. Weiss, A.K.; Carstens, T.V.: ¬The year's work in cataloging, 1999 (2001) 0.01
    0.0075104274 = product of:
      0.052572988 = sum of:
        0.037363302 = weight(_text_:web in 6084) [ClassicSimilarity], result of:
          0.037363302 = score(doc=6084,freq=4.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.35694647 = fieldWeight in 6084, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=6084)
        0.015209687 = product of:
          0.030419374 = sum of:
            0.030419374 = weight(_text_:22 in 6084) [ClassicSimilarity], result of:
              0.030419374 = score(doc=6084,freq=2.0), product of:
                0.11231873 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0320743 = queryNorm
                0.2708308 = fieldWeight in 6084, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=6084)
          0.5 = coord(1/2)
      0.14285715 = coord(2/14)
    
    Abstract
    The challenge of cataloging Web sites and electronic resources was the most important issue facing the cataloging world in the last year. This article reviews attempts to analyze and revise the cataloging code in view of the new electronic environment. The difficulties of applying traditional library cataloging standards to Web resources has led some to favor metadata as the best means of providing access to these materials. The appropriate education and training for library cataloging personnel remains crucial during this transitional period. Articles on user understanding of Library of Congress subject headings and on cataloging practice are also reviewed.
    Date
    10. 9.2000 17:38:22
  7. Trends der Informationsgesellschaft : eine annotierte Bibliographie zur Zukunft der Medienkommunikation (1995) 0.01
    0.0067287693 = product of:
      0.094202764 = sum of:
        0.094202764 = weight(_text_:medien in 2954) [ClassicSimilarity], result of:
          0.094202764 = score(doc=2954,freq=2.0), product of:
            0.15096188 = queryWeight, product of:
              4.7066307 = idf(docFreq=1085, maxDocs=44218)
              0.0320743 = queryNorm
            0.62401694 = fieldWeight in 2954, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.7066307 = idf(docFreq=1085, maxDocs=44218)
              0.09375 = fieldNorm(doc=2954)
      0.071428575 = coord(1/14)
    
    Series
    agenda Medien; 1
  8. Corbett, L.E.: Serials: review of the literature 2000-2003 (2006) 0.01
    0.0066850763 = product of:
      0.046795532 = sum of:
        0.035931468 = weight(_text_:open in 1088) [ClassicSimilarity], result of:
          0.035931468 = score(doc=1088,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.24876907 = fieldWeight in 1088, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1088)
        0.010864062 = product of:
          0.021728124 = sum of:
            0.021728124 = weight(_text_:22 in 1088) [ClassicSimilarity], result of:
              0.021728124 = score(doc=1088,freq=2.0), product of:
                0.11231873 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0320743 = queryNorm
                0.19345059 = fieldWeight in 1088, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1088)
          0.5 = coord(1/2)
      0.14285715 = coord(2/14)
    
    Abstract
    The topic of electronic journals (e-journals) dominated the serials literature from 2000 to 2003. This review is limited to the events and issues within the broad topics of cost, management, and archiving. Coverage of cost includes such initiatives as PEAK, JACC, BioMed Central, SPARC, open access, the "Big Deal," and "going e-only." Librarians combated the continued price increase trend for journals, fueled in part by publisher mergers, with the economies found with bundled packages and consortial subscriptions. Serials management topics include usage statistics; core title lists; staffing needs; the "A-Z list" and other services from such companies as Serials Solutions; "deep linking"; link resolvers such as SFX; development of standards or guidelines, such as COUNTER and ERMI; tracking of license terms; vendor mergers; and the demise of integrated library systems and a subscription agent's bankruptcy. Librarians archived print volumes in storage facilities due to space shortages. Librarians and publishers struggled with electronic archiving concepts, discussing questions of who, where, and how. Projects such as LOCKSS tested potential solutions, but missing online content due to the Tasini court case and retractions posed more archiving difficulties. The serials literature captured much of the upheaval resulting from the rapid pace of changes, many linked to the advent of e-journals.
    Date
    10. 9.2000 17:38:22
  9. Efthimiadis, E.N.; Neilson, C.: ¬A classified bibliography on online public access catalogues (1989) 0.01
    0.006221119 = product of:
      0.08709566 = sum of:
        0.08709566 = weight(_text_:source in 509) [ClassicSimilarity], result of:
          0.08709566 = score(doc=509,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.5477371 = fieldWeight in 509, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.078125 = fieldNorm(doc=509)
      0.071428575 = coord(1/14)
    
    Abstract
    Bibliography on all aspects of online public access catalogues, covering the period to 1988. Citations are listed in subject order supplemented by an author index. A description of the subject coverage, source of citations and how to use the bibliography is given in the introduction.
  10. Chambers, S.; Myall, C.: Cataloging and classification : review of the literature 2007-8 (2010) 0.01
    0.0059470762 = product of:
      0.04162953 = sum of:
        0.026419844 = weight(_text_:web in 4309) [ClassicSimilarity], result of:
          0.026419844 = score(doc=4309,freq=2.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.25239927 = fieldWeight in 4309, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4309)
        0.015209687 = product of:
          0.030419374 = sum of:
            0.030419374 = weight(_text_:22 in 4309) [ClassicSimilarity], result of:
              0.030419374 = score(doc=4309,freq=2.0), product of:
                0.11231873 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0320743 = queryNorm
                0.2708308 = fieldWeight in 4309, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4309)
          0.5 = coord(1/2)
      0.14285715 = coord(2/14)
    
    Abstract
    This paper surveys library literature on cataloging and classification published in 2007-8, indicating its extent and range in terms of types of literature, major subject areas, and themes. The paper reviews pertinent literature in the following areas: the future of bibliographic control, general cataloging standards and texts, Functional Requirements for Bibliographic Records (FRBR), cataloging varied resources, metadata and cataloging in the Web world, classification and subject access, questions of diversity and diverse perspectives, additional reports of practice and research, catalogers' education and careers, keeping current through columns and blogs, and cataloging history.
    Date
    10. 9.2000 17:38:22
  11. Chen, H.; Chau, M.: Web mining : machine learning for Web applications (2003) 0.01
    0.0051151155 = product of:
      0.07161161 = sum of:
        0.07161161 = weight(_text_:web in 4242) [ClassicSimilarity], result of:
          0.07161161 = score(doc=4242,freq=20.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.6841342 = fieldWeight in 4242, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4242)
      0.071428575 = coord(1/14)
    
    Abstract
    With more than two billion pages created by millions of Web page authors and organizations, the World Wide Web is a tremendously rich knowledge base. The knowledge comes not only from the content of the pages themselves, but also from the unique characteristics of the Web, such as its hyperlink structure and its diversity of content and languages. Analysis of these characteristics often reveals interesting patterns and new knowledge. Such knowledge can be used to improve users' efficiency and effectiveness in searching for information an the Web, and also for applications unrelated to the Web, such as support for decision making or business management. The Web's size and its unstructured and dynamic content, as well as its multilingual nature, make the extraction of useful knowledge a challenging research problem. Furthermore, the Web generates a large amount of data in other formats that contain valuable information. For example, Web server logs' information about user access patterns can be used for information personalization or improving Web page design.
  12. Large, A.: Children, teenagers, and the Web (2004) 0.00
    0.0043134443 = product of:
      0.060388215 = sum of:
        0.060388215 = weight(_text_:web in 4154) [ClassicSimilarity], result of:
          0.060388215 = score(doc=4154,freq=2.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.5769126 = fieldWeight in 4154, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.125 = fieldNorm(doc=4154)
      0.071428575 = coord(1/14)
    
  13. Yang, K.: Information retrieval on the Web (2004) 0.00
    0.0043134443 = product of:
      0.060388215 = sum of:
        0.060388215 = weight(_text_:web in 4278) [ClassicSimilarity], result of:
          0.060388215 = score(doc=4278,freq=32.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.5769126 = fieldWeight in 4278, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=4278)
      0.071428575 = coord(1/14)
    
    Abstract
    How do we find information an the Web? Although information on the Web is distributed and decentralized, the Web can be viewed as a single, virtual document collection. In that regard, the fundamental questions and approaches of traditional information retrieval (IR) research (e.g., term weighting, query expansion) are likely to be relevant in Web document retrieval. Findings from traditional IR research, however, may not always be applicable in a Web setting. The Web document collection - massive in size and diverse in content, format, purpose, and quality - challenges the validity of previous research findings that are based an relatively small and homogeneous test collections. Moreover, some traditional IR approaches, although applicable in theory, may be impossible or impractical to implement in a Web setting. For instance, the size, distribution, and dynamic nature of Web information make it extremely difficult to construct a complete and up-to-date data representation of the kind required for a model IR system. To further complicate matters, information seeking on the Web is diverse in character and unpredictable in nature. Web searchers come from all walks of life and are motivated by many kinds of information needs. The wide range of experience, knowledge, motivation, and purpose means that searchers can express diverse types of information needs in a wide variety of ways with differing criteria for satisfying those needs. Conventional evaluation measures, such as precision and recall, may no longer be appropriate for Web IR, where a representative test collection is all but impossible to construct. Finding information on the Web creates many new challenges for, and exacerbates some old problems in, IR research. At the same time, the Web is rich in new types of information not present in most IR test collections. Hyperlinks, usage statistics, document markup tags, and collections of topic hierarchies such as Yahoo! (http://www.yahoo.com) present an opportunity to leverage Web-specific document characteristics in novel ways that go beyond the term-based retrieval framework of traditional IR. Consequently, researchers in Web IR have reexamined the findings from traditional IR research.
  14. Jascó, P.: Content evaluation of databases (1997) 0.00
    0.003732671 = product of:
      0.052257393 = sum of:
        0.052257393 = weight(_text_:source in 2299) [ClassicSimilarity], result of:
          0.052257393 = score(doc=2299,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.32864225 = fieldWeight in 2299, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.046875 = fieldNorm(doc=2299)
      0.071428575 = coord(1/14)
    
    Abstract
    State of the art review of the criteria used to judge and evaluate the quality of databases, including: content, ease of use, accessibility, customer support, documentation, and value to cost ratio. Concludes that the principle factor governing quality is content, defined by the scope and the coverage of the database and its currency, accuracy, consistency and completeness. Scope is determined by its composition and coverage, including time period, number of periodicals and other primary sources, number of articles and geographic and language distribution. Currency is measured by the time lag between publication of the primary source and availability of the corresponding records in the database. Accuracy is governed by the extent to which the records are free from errors of all types. Consistency depends on the extent to which records within the database follow the same rules. Completeness is measured by the consistency with which applicable data elements are assigned to all the records in the database. Reviews the major contributions to the literature in the field and summarizes the background of milestone studies
  15. Davenport, E.; Hall, H.: Organizational Knowledge and Communities of Practice (2002) 0.00
    0.003732671 = product of:
      0.052257393 = sum of:
        0.052257393 = weight(_text_:source in 4293) [ClassicSimilarity], result of:
          0.052257393 = score(doc=4293,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.32864225 = fieldWeight in 4293, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.046875 = fieldNorm(doc=4293)
      0.071428575 = coord(1/14)
    
    Abstract
    A community of practice has recently been defined as "a flexible group of professionals, informally bound by common interests, who interact through interdependent tasks guided by a common purpose thereby embodying a store of common knowledge" (Jubert, 1999, p. 166). The association of communities of practice with the production of collective knowledge has long been recognized, and they have been objects of study for a number of decades in the context of professional communication, particularly communication in science (Abbott, 1988; Bazerman & Paradis, 1991). Recently, however, they have been invoked in the domain of organization studies as sites where people learn and share insights. If, as Stinchcombe suggests, an organization is "a set of stable social relations, dehberately created, with the explicit intention of continuously accomplishing some specific goals or purposes" (Stinchcombe, 1965, p. 142), where does this "flexible" and "embodied" source of knowledge fit? Can communities of practice be harnessed, engineered, and managed like other organizational groups, or does their strength lie in the fact that they operate outside the stable and persistent social relations that characterize the organization?
  16. Benoit, G.: Data mining (2002) 0.00
    0.003732671 = product of:
      0.052257393 = sum of:
        0.052257393 = weight(_text_:source in 4296) [ClassicSimilarity], result of:
          0.052257393 = score(doc=4296,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.32864225 = fieldWeight in 4296, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.046875 = fieldNorm(doc=4296)
      0.071428575 = coord(1/14)
    
    Abstract
    Data mining (DM) is a multistaged process of extracting previously unanticipated knowledge from large databases, and applying the results to decision making. Data mining tools detect patterns from the data and infer associations and rules from them. The extracted information may then be applied to prediction or classification models by identifying relations within the data records or between databases. Those patterns and rules can then guide decision making and forecast the effects of those decisions. However, this definition may be applied equally to "knowledge discovery in databases" (KDD). Indeed, in the recent literature of DM and KDD, a source of confusion has emerged, making it difficult to determine the exact parameters of both. KDD is sometimes viewed as the broader discipline, of which data mining is merely a component-specifically pattern extraction, evaluation, and cleansing methods (Raghavan, Deogun, & Sever, 1998, p. 397). Thurasingham (1999, p. 2) remarked that "knowledge discovery," "pattern discovery," "data dredging," "information extraction," and "knowledge mining" are all employed as synonyms for DM. Trybula, in his ARIST chapter an text mining, observed that the "existing work [in KDD] is confusing because the terminology is inconsistent and poorly defined.
  17. Rasmussen, E.M.: Indexing and retrieval for the Web (2002) 0.00
    0.0035305 = product of:
      0.049427 = sum of:
        0.049427 = weight(_text_:web in 4285) [ClassicSimilarity], result of:
          0.049427 = score(doc=4285,freq=28.0), product of:
            0.10467481 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0320743 = queryNorm
            0.47219574 = fieldWeight in 4285, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4285)
      0.071428575 = coord(1/14)
    
    Abstract
    The introduction and growth of the World Wide Web (WWW, or Web) have resulted in a profound change in the way individuals and organizations access information. In terms of volume, nature, and accessibility, the characteristics of electronic information are significantly different from those of even five or six years ago. Control of, and access to, this flood of information rely heavily an automated techniques for indexing and retrieval. According to Gudivada, Raghavan, Grosky, and Kasanagottu (1997, p. 58), "The ability to search and retrieve information from the Web efficiently and effectively is an enabling technology for realizing its full potential." Almost 93 percent of those surveyed consider the Web an "indispensable" Internet technology, second only to e-mail (Graphie, Visualization & Usability Center, 1998). Although there are other ways of locating information an the Web (browsing or following directory structures), 85 percent of users identify Web pages by means of a search engine (Graphie, Visualization & Usability Center, 1998). A more recent study conducted by the Stanford Institute for the Quantitative Study of Society confirms the finding that searching for information is second only to e-mail as an Internet activity (Nie & Ebring, 2000, online). In fact, Nie and Ebring conclude, "... the Internet today is a giant public library with a decidedly commercial tilt. The most widespread use of the Internet today is as an information search utility for products, travel, hobbies, and general information. Virtually all users interviewed responded that they engaged in one or more of these information gathering activities."
    Techniques for automated indexing and information retrieval (IR) have been developed, tested, and refined over the past 40 years, and are well documented (see, for example, Agosti & Smeaton, 1996; BaezaYates & Ribeiro-Neto, 1999a; Frakes & Baeza-Yates, 1992; Korfhage, 1997; Salton, 1989; Witten, Moffat, & Bell, 1999). With the introduction of the Web, and the capability to index and retrieve via search engines, these techniques have been extended to a new environment. They have been adopted, altered, and in some Gases extended to include new methods. "In short, search engines are indispensable for searching the Web, they employ a variety of relatively advanced IR techniques, and there are some peculiar aspects of search engines that make searching the Web different than more conventional information retrieval" (Gordon & Pathak, 1999, p. 145). The environment for information retrieval an the World Wide Web differs from that of "conventional" information retrieval in a number of fundamental ways. The collection is very large and changes continuously, with pages being added, deleted, and altered. Wide variability between the size, structure, focus, quality, and usefulness of documents makes Web documents much more heterogeneous than a typical electronic document collection. The wide variety of document types includes images, video, audio, and scripts, as well as many different document languages. Duplication of documents and sites is common. Documents are interconnected through networks of hyperlinks. Because of the size and dynamic nature of the Web, preprocessing all documents requires considerable resources and is often not feasible, certainly not an the frequent basis required to ensure currency. Query length is usually much shorter than in other environments-only a few words-and user behavior differs from that in other environments. These differences make the Web a novel environment for information retrieval (Baeza-Yates & Ribeiro-Neto, 1999b; Bharat & Henzinger, 1998; Huang, 2000).
  18. Singh, S. (Sewa); Singh, S. (Sukhbir): Colon Classification : a select bibliography (1992) 0.00
    0.0035191963 = product of:
      0.049268745 = sum of:
        0.049268745 = weight(_text_:source in 1479) [ClassicSimilarity], result of:
          0.049268745 = score(doc=1479,freq=4.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.3098469 = fieldWeight in 1479, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.03125 = fieldNorm(doc=1479)
      0.071428575 = coord(1/14)
    
    Abstract
    This being the Birth Centenary.Year of Dr. S. R. Ranganathan, it was deemed as a befitting tribute to the author of Colon Classifiction to bring together all that has been published thus far on it and complile it in the form of a bibliographies are an important source of information on a particular field of study and research. A research worker has to carry out the literature survey on the area of interest which sometimes consumes many precious man hours in order to collect some relevent citations. Thus bibliographies are a useful source in saving the precious time and energy of the reseach workers, and lead them to the primary sources for carryying out the research further to the destination.
  19. Shubert, S.B.: Critical views of LCSH - ten years later : a bibliographic essay (1992) 0.00
    0.0031105594 = product of:
      0.04354783 = sum of:
        0.04354783 = weight(_text_:source in 5737) [ClassicSimilarity], result of:
          0.04354783 = score(doc=5737,freq=2.0), product of:
            0.15900996 = queryWeight, product of:
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0320743 = queryNorm
            0.27386856 = fieldWeight in 5737, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.9575505 = idf(docFreq=844, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5737)
      0.071428575 = coord(1/14)
    
    Abstract
    Pauline Cochrane and Monika Kirtland's "Critical views of LCSH-Library of Congress Subject Headings: a bibliographic and bibliometric essay" published in 'Cataloging & classification quarterly 1(1982) S.71-93' has been widely cited as a source for discussion and complaints about LCSH. Cochrane and Kirtland cover the literature from 1944-1979. The present work provides a critique of the Cochrane/Kirtland study and a survey of the literature concerning the LCSHs during the 1980s. The classified bibliography is arranged according to the format of the Cochrane/Kirtland study to facilitate comparison. Criticism of LCSH reiterates the same points over and over again, whether it is sparseness, bias or currency of the subject headings. Significant trends which emerged in the 1980s include an increased emphasis on the use of LCSH as an online searching tool, concern for the syndetic structure of LCSH and the role of subdivisions, as well as repeated calls for the development of coherent standards to ensure LCSH evolves and is applied in a consistent manner
  20. Kling, R.: ¬The Internet and unrefereed scholarly publishing (2003) 0.00
    0.00307984 = product of:
      0.043117758 = sum of:
        0.043117758 = weight(_text_:open in 4272) [ClassicSimilarity], result of:
          0.043117758 = score(doc=4272,freq=2.0), product of:
            0.14443703 = queryWeight, product of:
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.0320743 = queryNorm
            0.2985229 = fieldWeight in 4272, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.5032015 = idf(docFreq=1330, maxDocs=44218)
              0.046875 = fieldNorm(doc=4272)
      0.071428575 = coord(1/14)
    
    Abstract
    In the early 1990s, much of the enthusiasm for the use of electronic media to enhance scholarly communication focused an electronic journals, especially electronic-only, (pure) e journals (see for example, Peek & Newby's [1996] anthology). Much of the systematic research an the use of electronic media to enhance scholarly communication also focused an electronic journals. However, by the late 1990s, numerous scientific publishers had transformed their paper journals (p journals) into paper and electronic journals (p-e journals) and sold them via subscription models that did not provide the significant costs savings, speed of access, or breadth of audience that pure e -journal advocates had expected (Okerson, 1996). In 2001, a group of senior life scientists led a campaign to have publishers make their journals freely available online six months after publication (Russo, 2001). The campaign leaders, using the name "Public Library of Science," asked scientists to boycott journals that did not comply with these demands for open access. Although the proposal was discussed in scientific magazines and conferences, it apparently did not persuade any journal publishers to comply (Young, 2002). Most productive scientists, who work for major universities and research institutes

Years

Languages

  • e 54
  • d 1
  • More… Less…

Types

  • a 48
  • b 14
  • m 3
  • el 2
  • r 1
  • More… Less…