Search (36 results, page 1 of 2)

  • × language_ss:"e"
  • × theme_ss:"Internet"
  • × type_ss:"m"
  1. Gundavaram, S.: CGI programming on the World Wide Web (1996) 0.04
    0.042530965 = product of:
      0.17012386 = sum of:
        0.17012386 = product of:
          0.34024772 = sum of:
            0.34024772 = weight(_text_:programming in 5948) [ClassicSimilarity], result of:
              0.34024772 = score(doc=5948,freq=2.0), product of:
                0.29361802 = queryWeight, product of:
                  6.5552235 = idf(docFreq=170, maxDocs=44218)
                  0.04479146 = queryNorm
                1.1588107 = fieldWeight in 5948, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  6.5552235 = idf(docFreq=170, maxDocs=44218)
                  0.125 = fieldNorm(doc=5948)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
  2. Segev, E.: Google and the digital divide : the bias of online knowledge (2010) 0.04
    0.035707813 = product of:
      0.14283125 = sum of:
        0.14283125 = weight(_text_:engines in 3079) [ClassicSimilarity], result of:
          0.14283125 = score(doc=3079,freq=10.0), product of:
            0.22757743 = queryWeight, product of:
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.04479146 = queryNorm
            0.62761605 = fieldWeight in 3079, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3079)
      0.25 = coord(1/4)
    
    Abstract
    Aimed at information and communication professionals, scholars and students, Google and the Digital Divide: The Biases of Online Knowledge provides invaluable insight into the significant role that search engines play in growing the digital divide between individuals, organizations, and states. With a specific focus on Google, author Elad Segev explains the concept of the digital divide and the effects that today's online environment has on knowledge bias, power, and control. Using innovative methods and research approaches, Segev compares the popular search queries in Google and Yahoo in the United States and other countries and analyzes the various biases in Google News and Google Earth. Google and the Digital Divide shows the many ways in which users manipulate Google's information across different countries, as well as dataset and classification systems, economic and political value indexes, specific search indexes, locality of use indexes, and much more. Segev presents important new social and political perspectives to illustrate the challenges brought about by search engines, and explains the resultant political, communicative, commercial, and international implications.
    Content
    Inhalt: Power, communication and the internet -- The structure and power of search engines -- Google and the politics of online searching -- Users and uses of Google's information -- Mass media channels and the world of Google News -- Google's global mapping
    LCSH
    Search engines
    Subject
    Search engines
  3. Bizer, C.; Mendes, P.N.; Jentzsch, A.: Topology of the Web of Data (2012) 0.03
    0.031292755 = product of:
      0.12517102 = sum of:
        0.12517102 = weight(_text_:engines in 425) [ClassicSimilarity], result of:
          0.12517102 = score(doc=425,freq=12.0), product of:
            0.22757743 = queryWeight, product of:
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.04479146 = queryNorm
            0.5500151 = fieldWeight in 425, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.03125 = fieldNorm(doc=425)
      0.25 = coord(1/4)
    
    Abstract
    The degree of structure of Web content is the determining factor for the types of functionality that search engines can provide. The more well structured the Web content is, the easier it is for search engines to understand Web content and provide advanced functionality, such as faceted filtering or the aggregation of content from multiple Web sites, based on this understanding. Today, most Web sites are generated from structured data that is stored in relational databases. Thus, it does not require too much extra effort for Web sites to publish this structured data directly on the Web in addition to HTML pages, and thus help search engines to understand Web content and provide improved functionality. An early approach to realize this idea and help search engines to understand Web content is Microformats, a technique for markingup structured data about specific types on entities-such as tags, blog posts, people, or reviews-within HTML pages. As Microformats are focused on a few entity types, the World Wide Web Consortium (W3C) started in 2004 to standardize RDFa as an alternative, more generic language for embedding any type of data into HTML pages. Today, major search engines such as Google, Yahoo, and Bing extract Microformat and RDFa data describing products, reviews, persons, events, and recipes from Web pages and use the extracted data to improve the user's search experience. The search engines have started to aggregate structured data from different Web sites and augment their search results with these aggregated information units in the form of rich snippets which combine, for instance, data This chapter gives an overview of the topology of the Web of Data that has been created by publishing data on the Web using the microformats RDFa, Microdata and Linked Data publishing techniques.
  4. Internet searching and indexing : the subject approach (2000) 0.03
    0.025550429 = product of:
      0.102201715 = sum of:
        0.102201715 = weight(_text_:engines in 1468) [ClassicSimilarity], result of:
          0.102201715 = score(doc=1468,freq=2.0), product of:
            0.22757743 = queryWeight, product of:
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.04479146 = queryNorm
            0.44908544 = fieldWeight in 1468, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.0625 = fieldNorm(doc=1468)
      0.25 = coord(1/4)
    
    Abstract
    This comprehensive volume offers usable information for people at all levels of Internet savvy. It can teach librarians, students, and patrons how to search the Internet more systematically. It also helps information professionals design more efficient, effective search engines and Web pages.
  5. Sherman, C.; Price, G.: ¬The invisible Web : uncovering information sources search engines can't see (2001) 0.02
    0.022583602 = product of:
      0.09033441 = sum of:
        0.09033441 = weight(_text_:engines in 62) [ClassicSimilarity], result of:
          0.09033441 = score(doc=62,freq=4.0), product of:
            0.22757743 = queryWeight, product of:
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.04479146 = queryNorm
            0.39693922 = fieldWeight in 62, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.0390625 = fieldNorm(doc=62)
      0.25 = coord(1/4)
    
    Abstract
    Enormous expanses of the Internet are unreachable with standard Web search engines. This book provides the key to finding these hidden resources by identifying how to uncover and use invisible Web resources. Mapping the invisible Web, when and how to use it, assessing the validity of the information, and the future of Web searching are topics covered in detail. Only 16 percent of Net-based information can be located using a general search engine. The other 84 percent is what is referred to as the invisible Web-made up of information stored in databases. Unlike pages on the visible Web, information in databases is generally inaccessible to the software spiders and crawlers that compile search engine indexes. As Web technology improves, more and more information is being stored in databases that feed into dynamically generated Web pages. The tips provided in this resource will ensure that those databases are exposed and Net-based research will be conducted in the most thorough and effective manner. Discusses the use of online information resources and problems caused by dynamically generated Web pages, paying special attention to information mapping, assessing the validity of information, and the future of Web searching.
  6. Bates, C.: Web programming : building Internet applications (2000) 0.02
    0.021265483 = product of:
      0.08506193 = sum of:
        0.08506193 = product of:
          0.17012386 = sum of:
            0.17012386 = weight(_text_:programming in 5) [ClassicSimilarity], result of:
              0.17012386 = score(doc=5,freq=2.0), product of:
                0.29361802 = queryWeight, product of:
                  6.5552235 = idf(docFreq=170, maxDocs=44218)
                  0.04479146 = queryNorm
                0.57940537 = fieldWeight in 5, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  6.5552235 = idf(docFreq=170, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
  7. Mesaric, G.: Black magic Web : using the new World Wide Web technologies (1997) 0.02
    0.018607298 = product of:
      0.07442919 = sum of:
        0.07442919 = product of:
          0.14885838 = sum of:
            0.14885838 = weight(_text_:programming in 5395) [ClassicSimilarity], result of:
              0.14885838 = score(doc=5395,freq=2.0), product of:
                0.29361802 = queryWeight, product of:
                  6.5552235 = idf(docFreq=170, maxDocs=44218)
                  0.04479146 = queryNorm
                0.5069797 = fieldWeight in 5395, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  6.5552235 = idf(docFreq=170, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5395)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    A spell book for mastering the WWW. Step by step, readers are taught how to use the Internet efficiently, and, in particular, how to publish information on the Web. After completing this book, the reader will be able, for example, to create stunning Web home pages and even take the first steps in Java programming. There are many books on the net bound to either one product, e.g. Mosaic, or to one topic, HTML. Mesaric, on the other hand, covers all WWW-related topics and all products which are in popular use. The magic revealed includes common data formats (HTML, PDF, JPEG, GIF), an introduction to Java and Java Srcipt, emerging technologies such as Hyper-G, hypermedia authoring with Haemony and Amadeus, information about ATM, TeleScript, and much more
  8. Fisher, Y.: Spinning the Web : a guide to serving information on the World Wide Web (1996) 0.02
    0.018607298 = product of:
      0.07442919 = sum of:
        0.07442919 = product of:
          0.14885838 = sum of:
            0.14885838 = weight(_text_:programming in 5946) [ClassicSimilarity], result of:
              0.14885838 = score(doc=5946,freq=2.0), product of:
                0.29361802 = queryWeight, product of:
                  6.5552235 = idf(docFreq=170, maxDocs=44218)
                  0.04479146 = queryNorm
                0.5069797 = fieldWeight in 5946, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  6.5552235 = idf(docFreq=170, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5946)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    Most books on the Internet describe it from the user's end. This one, however, is unique in its focus on serving information on the WWW. It presents everything from the basics to advanced techniques and will thus prove invaluable to site administrators and developers. The author - an expert developer and researcher at UCSD - covers such topics as HTML 3.0, serving documents, interfaces, WWW utilities and browsers such as Netscape. Fisher also includes an introduction to programming with JAVA and JAVA sript, as well as the complete VRML 1.0 specification
  9. Social information retrieval systems : emerging technologies and applications for searching the Web effectively (2008) 0.02
    0.018066881 = product of:
      0.072267525 = sum of:
        0.072267525 = weight(_text_:engines in 4127) [ClassicSimilarity], result of:
          0.072267525 = score(doc=4127,freq=4.0), product of:
            0.22757743 = queryWeight, product of:
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.04479146 = queryNorm
            0.31755137 = fieldWeight in 4127, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.03125 = fieldNorm(doc=4127)
      0.25 = coord(1/4)
    
    LCSH
    Web search engines
    Subject
    Web search engines
  10. Rogers, R.: Digital methods (2013) 0.02
    0.018066881 = product of:
      0.072267525 = sum of:
        0.072267525 = weight(_text_:engines in 2354) [ClassicSimilarity], result of:
          0.072267525 = score(doc=2354,freq=4.0), product of:
            0.22757743 = queryWeight, product of:
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.04479146 = queryNorm
            0.31755137 = fieldWeight in 2354, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.03125 = fieldNorm(doc=2354)
      0.25 = coord(1/4)
    
    LCSH
    Web search engines
    Subject
    Web search engines
  11. Stuart, D.: Web metrics for library and information professionals (2014) 0.02
    0.01580852 = product of:
      0.06323408 = sum of:
        0.06323408 = weight(_text_:engines in 2274) [ClassicSimilarity], result of:
          0.06323408 = score(doc=2274,freq=4.0), product of:
            0.22757743 = queryWeight, product of:
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.04479146 = queryNorm
            0.27785745 = fieldWeight in 2274, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.02734375 = fieldNorm(doc=2274)
      0.25 = coord(1/4)
    
    Content
    1. Introduction. MetricsIndicators -- Web metrics and Ranganathan's laws of library science -- Web metrics for the library and information professional -- The aim of this book -- The structure of the rest of this book -- 2. Bibliometrics, webometrics and web metrics. Web metrics -- Information science metrics -- Web analytics -- Relational and evaluative metrics -- Evaluative web metrics -- Relational web metrics -- Validating the results -- 3. Data collection tools. The anatomy of a URL, web links and the structure of the web -- Search engines 1.0 -- Web crawlers -- Search engines 2.0 -- Post search engine 2.0: fragmentation -- 4. Evaluating impact on the web. Websites -- Blogs -- Wikis -- Internal metrics -- External metrics -- A systematic approach to content analysis -- 5. Evaluating social media impact. Aspects of social network sites -- Typology of social network sites -- Research and tools for specific sites and services -- Other social network sites -- URL shorteners: web analytic links on any site -- General social media impact -- Sentiment analysis -- 6. Investigating relationships between actors. Social network analysis methods -- Sources for relational network analysis -- 7. Exploring traditional publications in a new environment. More bibliographic items -- Full text analysis -- Greater context -- 8. Web metrics and the web of data. The web of data -- Building the semantic web -- Implications of the web of data for web metrics -- Investigating the web of data today -- SPARQL -- Sindice -- LDSpider: an RDF web crawler -- 9. The future of web metrics and the library and information professional. How far we have come -- The future of web metrics -- The future of the library and information professional and web metrics.
  12. Schwartz, C.: Sorting out the Web : approaches to subject access (2001) 0.01
    0.013829577 = product of:
      0.055318307 = sum of:
        0.055318307 = weight(_text_:engines in 2050) [ClassicSimilarity], result of:
          0.055318307 = score(doc=2050,freq=6.0), product of:
            0.22757743 = queryWeight, product of:
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.04479146 = queryNorm
            0.24307466 = fieldWeight in 2050, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.01953125 = fieldNorm(doc=2050)
      0.25 = coord(1/4)
    
    Footnote
    Rez. in: KO 50(2003) no.1, S.45-46 (L.M. Given): "In her own preface to this work, the author notes her lifelong fascination with classification and order, as well as her more recent captivation with the Internet - a place of "chaos in need of organization" (xi). Sorting out the Web examines current efforts to organize the Web and is well-informed by the author's academic and professional expertise in information organization, information retrieval, and Web development. Although the book's level and tone are particularly relevant to a student audience (or others interested in Web-based subject access at an introductory level), it will also appeal to information professionals developing subject access systems across a range of information contexts. There are six chapters in the book, each describing and analyzing one core concept related to the organization of Web content. All topics are presented in a manner ideal for newcomers to the area, with clear definitions, examples, and visuals that illustrate the principles under discussion. The first chapter provides a brief introduction to developments in information technology, including an historical overview of information services, users' needs, and libraries' responses to the Internet. Chapter two introduces metadata, including core concepts and metadata formats. Throughout this chapter the author presents a number of figures that aptly illustrate the application of metadata in HTML, SGML, and MARC record environments, and the use of metadata tools (e.g., XML, RDF). Chapter three begins with an overview of classification theory and specific schemes, but the author devotes most of the discussion to the application of classification systems in the Web environment (e.g., Dewey, LCC, UDC). Web screen captures illustrate the use of these schemes for information sources posted to sites around the world. The chapter closes with a discussion of the future of classification; this is a particularly useful section as the author presents a listing of core journal and conference venues where new approaches to Web classification are explored. In chapter four, the author extends the discussion of classification to the use of controlled vocabularies. As in the first few chapters, the author first presents core background material, including reasons to use controlled vocabularies and the differences between preand post-coordinate indexing, and then discusses the application of specific vocabularies in the Web environment (e.g., Infomine's use of LCSH). The final section of the chapter explores failure in subject searching and the limitations of controlled vocabularies for the Web. Chapter five discusses one of the most common and fast-growing topics related to subject access an the Web: search engines. The author presents a clear definition of the term that encompasses classified search lists (e.g., Yahoo) and query-based engines (e.g., Alta Vista). In addition to historical background an the development of search engines, Schwartz also examines search service types, features, results, and system performance.
  13. Research and advanced technology for digital libraries : 9th European conference, ECDL 2005, Vienna, Austria, September 18 - 23, 2005 ; proceedings (2005) 0.01
    0.012775214 = product of:
      0.051100858 = sum of:
        0.051100858 = weight(_text_:engines in 2423) [ClassicSimilarity], result of:
          0.051100858 = score(doc=2423,freq=2.0), product of:
            0.22757743 = queryWeight, product of:
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.04479146 = queryNorm
            0.22454272 = fieldWeight in 2423, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.03125 = fieldNorm(doc=2423)
      0.25 = coord(1/4)
    
    Content
    Inhalt u.a.: - Digital Library Models and Architectures - Multimedia and Hypermedia Digital Libraries - XML - Building Digital Libraries - User Studies - Digital Preservation - Metadata - Digital Libraries and e-Learning - Text Classification in Digital Libraries - Searching - - Focused Crawling Using Latent Semantic Indexing - An Application for Vertical Search Engines / George Almpanidis, Constantine Kotropoulos, Ioannis Pitas - - Active Support for Query Formulation in Virtual Digital Libraries: A Case Study with DAFFODIL / Andre Schaefer, Matthias Jordan, Claus-Peter Klas, Norbert Fuhr - - Expression of Z39.50 Supported Search Capabilities by Applying Formal Descriptions / Michalis Sfakakis, Sarantos Kapidakis - Text Digital Libraries
  14. Bruce, H.: ¬The user's view of the Internet (2002) 0.01
    0.011857145 = product of:
      0.02371429 = sum of:
        0.01916282 = weight(_text_:engines in 4344) [ClassicSimilarity], result of:
          0.01916282 = score(doc=4344,freq=2.0), product of:
            0.22757743 = queryWeight, product of:
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.04479146 = queryNorm
            0.08420352 = fieldWeight in 4344, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.01171875 = fieldNorm(doc=4344)
        0.0045514684 = product of:
          0.009102937 = sum of:
            0.009102937 = weight(_text_:22 in 4344) [ClassicSimilarity], result of:
              0.009102937 = score(doc=4344,freq=2.0), product of:
                0.15685207 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04479146 = queryNorm
                0.058035173 = fieldWeight in 4344, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.01171875 = fieldNorm(doc=4344)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Footnote
    Chapter 2 (Technology and People) focuses an several theories of technological acceptance and diffusion. Unfortunately, Bruce's presentation is somewhat confusing as he moves from one theory to next, never quite connecting them into a logical sequence or coherent whole. Two theories are of particular interest to Bruce: the Theory of Diffusion of Innovations and the Theory of Planned Behavior. The Theory of Diffusion of Innovations is an "information-centric view of technology acceptance" in which technology adopters are placed in the information flows of society from which they learn about innovations and "drive innovation adoption decisions" (p. 20). The Theory of Planned Behavior maintains that the "performance of a behavior is a joint function of intentions and perceived behavioral control" (i.e., how muck control a person thinks they have) (pp. 22-23). Bruce combines these two theories to form the basis for the Technology Acceptance Model. This model posits that "an individual's acceptance of information technology is based an beliefs, attitudes, intentions, and behaviors" (p. 24). In all these theories and models echoes a recurring theme: "individual perceptions of the innovation or technology are critical" in terms of both its characteristics and its use (pp. 24-25). From these, in turn, Bruce derives a predictive theory of the role personal perceptions play in technology adoption: Personal Innovativeness of Information Technology Adoption (PIITA). Personal inventiveness is defined as "the willingness of an individual to try out any new information technology" (p. 26). In general, the PIITA theory predicts that information technology will be adopted by individuals that have a greater exposure to mass media, rely less an the evaluation of information technology by others, exhibit a greater ability to cope with uncertainty and take risks, and requires a less positive perception of an information technology prior to its adoption. Chapter 3 (A Focus an Usings) introduces the User-Centered Paradigm (UCP). The UCP is characteristic of the shift of emphasis from technology to users as the driving force behind technology and research agendas for Internet development [for a dissenting view, see Andrew Dillion's (2003) challenge to the utility of user-centerness for design guidance]. It entails the "broad acceptance of the user-oriented perspective across a range of disciplines and professional fields," such as business, education, cognitive engineering, and information science (p. 34).
    Bruce points out that Dervin is also concerned about how "we look at the world" in terms of "information needs and seeking" (p.60). She maintains that information scientists traditionally view information seeking and needs in terms of "contexts, users, and systems." Dervin questions whether or not, from a user's point of view, these three "points of interest" even exist. Rather it is the "micromoments of human usings" [emphasis original], and the "world viewings, seekings, and valuings" that comprise them that are real (p. 60). Using his metatheory, Bruce represents the Internet, the "object" of study, as a "chain of transformations made up of the micromoments of human usings" (p. 60). The Internet then is a "composite of usings" that, through research and study, is continuously reduced in complexity while its "essence" and "explanation" are amplified (p. 60). Bruce plans to use the Metatheory of Circulating Usings as an analytical "lens" to "tease out a characterization of the micromoments of Internet usings" from previous research an the Internet thereby exposing "the user's view of the Internet" (pp. 60-61). In Chapter 4 (Users of the Internet), Bruce presents the research data for the study. He begins with an explanation of the limits of the data, and to a certain extent, the study itself. The perspective is that of the Internet user, with a focus an use, not nonuse, thereby exluding issues such as the digital divide and universal service. The research is limited to Internet users "in modern economies around the world" (p. 60). The data is a synthesis of research from many disciplines, but mainly from those "associated with the information field" with its traditional focus an users, systems, and context rather than usings (p. 70). Bruce then presents an extensive summary of the research results from a massive literature review of available Internet studies. He examines the research for each study group in order of the amount of data available, starting with the most studied group professional users ("academics, librarians, and teachers") followed by "the younger generation" ("College students, youths, and young adults"), users of e-government information and e-business services, and ending with the general public (the least studied group) (p. 70). Bruce does a masterful job of condensing and summarizing a vast amount of research data in 49 pages. Although there is too muck to recapitulate here, one can get a sense of the results by looking at the areas of data examined for one of the study groups: academic Internet users. There is data an their frequency of use, reasons for nonuse, length of use, specific types of use (e.g., research, teaching, administration), use of discussion lists, use of e-journals, use of Web browsers and search engines, how academics learn to use web tools and services (mainly by self-instruction), factors affecting use, and information seeking habits. Bruce's goal in presenting all this research data is to provide "the foundation for constructs of the Internet that can inform stakeholders who will play a role in determining how the Internet will develop" (p. 129). These constructs are presented in Chapter 5.
  15. Clyde, L.A.: Weblogs and libraries (2004) 0.01
    0.011178312 = product of:
      0.044713248 = sum of:
        0.044713248 = weight(_text_:engines in 4496) [ClassicSimilarity], result of:
          0.044713248 = score(doc=4496,freq=2.0), product of:
            0.22757743 = queryWeight, product of:
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.04479146 = queryNorm
            0.19647488 = fieldWeight in 4496, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4496)
      0.25 = coord(1/4)
    
    Abstract
    This book discusses the topic of 'weblogs and libraries' from two main perspectives: weblogs as sources of information for libraries and librarians; and weblogs as tools that libraries can use to promote their services and to provide a means of communication with their clients. It begins with an overview of the whole weblog and blogging phenomenon and traces its development over the last six years. The many different kinds of weblogs are outlined (including personal weblogs, community weblogs, multimedia weblogs). The problem of locating weblogs is addressed through a discussion of weblog directories, search engines and other finding tools. Chapters include using weblogs as sources of information in the library or information service, the options for creating a weblog, and managing the library's own weblog.
  16. Rogers, R.: Information politics on the Web (2004) 0.01
    0.011063661 = product of:
      0.044254646 = sum of:
        0.044254646 = weight(_text_:engines in 442) [ClassicSimilarity], result of:
          0.044254646 = score(doc=442,freq=6.0), product of:
            0.22757743 = queryWeight, product of:
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.04479146 = queryNorm
            0.19445972 = fieldWeight in 442, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              5.080822 = idf(docFreq=746, maxDocs=44218)
              0.015625 = fieldNorm(doc=442)
      0.25 = coord(1/4)
    
    Footnote
    Rez. in: JASIST 58(2007) no.4, S.608-609 (K.D. Desouza): "Richard Rogers explores the distinctiveness of the World Wide Web as a politically contested space where information searchers may encounter multiple explanations of reality. Sources of information on the Web are in constant competition with each other for attention. The attention a source receives will determine its prominence, the ability to be a provider of leading information, and its inclusion in authoritative spaces. Rogers explores the politics behind evaluating sources that are collected and housed on authoritative spaces. Information politics on the Web can be looked at in terms of frontend or back-end politics. Front-end politics is concerned with whether sources on the Web pay attention to principles of inclusivity, fairness, and scope of representation in how information is presented, while back-end politics examines the logic behind how search engines or portals select and index information. Concerning front-end politics, Rogers questions the various versions of reality one can derive from examining information on the Web, especially when issues of information inclusivity and scope of representation are toiled with. In addition, Rogers is concerned with how back-end politics are being controlled by dominant forces of the market (i.e., the more an organization is willing to pay, the greater will be the site's visibility and prominence in authoritative spaces), regardless of whether the information presented on the site justifies such a placement. In the book, Rogers illustrates the issues involved in back-end and front-end politics (though heavily slanted on front-end politics) using vivid cases, all of which are derived from his own research. The main thrust is the exploration of how various "information instruments," defined as "a digital and analytical means of recording (capturing) and subsequently reading indications of states of defined information streams (p. 19)," help capture the politics of the Web. Rogers employs four specific instruments (Lay Decision Support System, Issue Barometer, Web Issue Index of Civil Society, and Election Issue Tracker), which are covered in detail in core chapters of the book (Chapter 2-Chapter 5). The book is comprised of six chapters, with Chapter 1 being the traditional introduction and Chapter 6 being a summary of the major concepts discussed.
    LCSH
    Web search engines / Political aspects
    Subject
    Web search engines / Political aspects
  17. Orenstein, R.M.: Fulltext sources online (1997) 0.01
    0.010620093 = product of:
      0.042480372 = sum of:
        0.042480372 = product of:
          0.084960744 = sum of:
            0.084960744 = weight(_text_:22 in 2677) [ClassicSimilarity], result of:
              0.084960744 = score(doc=2677,freq=2.0), product of:
                0.15685207 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04479146 = queryNorm
                0.5416616 = fieldWeight in 2677, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=2677)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Footnote
    Rez. in: Online 22(1998) no.1, S.93-94 (J. Alita)
  18. Tilman, H.N.: Internet tools of the profession : a guide for information professionals (1997) 0.01
    0.010620093 = product of:
      0.042480372 = sum of:
        0.042480372 = product of:
          0.084960744 = sum of:
            0.084960744 = weight(_text_:22 in 6620) [ClassicSimilarity], result of:
              0.084960744 = score(doc=6620,freq=2.0), product of:
                0.15685207 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04479146 = queryNorm
                0.5416616 = fieldWeight in 6620, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6620)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Footnote
    Rez. in: Online 22(1998) no.2, S.92 (C.A. Murchie)
  19. Gilster, P.: Digital literacy (1997) 0.01
    0.010620093 = product of:
      0.042480372 = sum of:
        0.042480372 = product of:
          0.084960744 = sum of:
            0.084960744 = weight(_text_:22 in 3340) [ClassicSimilarity], result of:
              0.084960744 = score(doc=3340,freq=2.0), product of:
                0.15685207 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04479146 = queryNorm
                0.5416616 = fieldWeight in 3340, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3340)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Footnote
    Rez. in: Online and CD-ROM review 22(1998) no.5, S.350-351 (P. Bradley)
  20. Cady, G.H.; McGregor, P.: Mastering the Internet (1996) 0.01
    0.009102937 = product of:
      0.036411747 = sum of:
        0.036411747 = product of:
          0.072823495 = sum of:
            0.072823495 = weight(_text_:22 in 5716) [ClassicSimilarity], result of:
              0.072823495 = score(doc=5716,freq=2.0), product of:
                0.15685207 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04479146 = queryNorm
                0.46428138 = fieldWeight in 5716, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=5716)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    13. 7.1998 19:21:22

Subjects

Classifications