Search (67 results, page 1 of 4)

  • × theme_ss:"Literaturübersicht"
  1. Thelwall, M.; Vaughan, L.; Björneborn, L.: Webometrics (2004) 0.02
    0.02308332 = product of:
      0.115416594 = sum of:
        0.050501734 = weight(_text_:web in 4279) [ClassicSimilarity], result of:
          0.050501734 = score(doc=4279,freq=18.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.5408555 = fieldWeight in 4279, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4279)
        0.06491486 = weight(_text_:log in 4279) [ClassicSimilarity], result of:
          0.06491486 = score(doc=4279,freq=2.0), product of:
            0.18335998 = queryWeight, product of:
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.028611459 = queryNorm
            0.3540296 = fieldWeight in 4279, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4279)
      0.2 = coord(2/10)
    
    Abstract
    Webometrics, the quantitative study of Web-related phenomena, emerged from the realization that methods originally designed for bibliometric analysis of scientific journal article citation patterns could be applied to the Web, with commercial search engines providing the raw data. Almind and Ingwersen (1997) defined the field and gave it its name. Other pioneers included Rodriguez Gairin (1997) and Aguillo (1998). Larson (1996) undertook exploratory link structure analysis, as did Rousseau (1997). Webometrics encompasses research from fields beyond information science such as communication studies, statistical physics, and computer science. In this review we concentrate on link analysis, but also cover other aspects of webometrics, including Web log fle analysis. One theme that runs through this chapter is the messiness of Web data and the need for data cleansing heuristics. The uncontrolled Web creates numerous problems in the interpretation of results, for instance, from the automatic creation or replication of links. The loose connection between top-level domain specifications (e.g., com, edu, and org) and their actual content is also a frustrating problem. For example, many .com sites contain noncommercial content, although com is ostensibly the main commercial top-level domain. Indeed, a skeptical researcher could claim that obstacles of this kind are so great that all Web analyses lack value. As will be seen, one response to this view, a view shared by critics of evaluative bibliometrics, is to demonstrate that Web data correlate significantly with some non-Web data in order to prove that the Web data are not wholly random. A practical response has been to develop increasingly sophisticated data cleansing techniques and multiple data analysis methods.
  2. Bar-Ilan, J.: ¬The use of Web search engines in information science research (2003) 0.01
    0.014964284 = product of:
      0.07482142 = sum of:
        0.06699813 = weight(_text_:web in 4271) [ClassicSimilarity], result of:
          0.06699813 = score(doc=4271,freq=22.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.717526 = fieldWeight in 4271, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4271)
        0.007823291 = product of:
          0.023469873 = sum of:
            0.023469873 = weight(_text_:29 in 4271) [ClassicSimilarity], result of:
              0.023469873 = score(doc=4271,freq=2.0), product of:
                0.10064617 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.028611459 = queryNorm
                0.23319192 = fieldWeight in 4271, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4271)
          0.33333334 = coord(1/3)
      0.2 = coord(2/10)
    
    Abstract
    The World Wide Web was created in 1989, but it has already become a major information channel and source, influencing our everyday lives, commercial transactions, and scientific communication, to mention just a few areas. The seventeenth-century philosopher Descartes proclaimed, "I think, therefore I am" (cogito, ergo sum). Today the Web is such an integral part of our lives that we could rephrase Descartes' statement as "I have a Web presence, therefore I am." Because many people, companies, and organizations take this notion seriously, in addition to more substantial reasons for publishing information an the Web, the number of Web pages is in the billions and growing constantly. However, it is not sufficient to have a Web presence; tools that enable users to locate Web pages are needed as well. The major tools for discovering and locating information an the Web are search engines. This review discusses the use of Web search engines in information science research. Before going into detail, we should define the terms "information science," "Web search engine," and "use" in the context of this review.
    Date
    23.10.2005 18:29:16
  3. Weiss, A.K.; Carstens, T.V.: ¬The year's work in cataloging, 1999 (2001) 0.01
    0.008474903 = product of:
      0.042374514 = sum of:
        0.033329446 = weight(_text_:web in 6084) [ClassicSimilarity], result of:
          0.033329446 = score(doc=6084,freq=4.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.35694647 = fieldWeight in 6084, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=6084)
        0.009045068 = product of:
          0.027135205 = sum of:
            0.027135205 = weight(_text_:22 in 6084) [ClassicSimilarity], result of:
              0.027135205 = score(doc=6084,freq=2.0), product of:
                0.10019246 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.028611459 = queryNorm
                0.2708308 = fieldWeight in 6084, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=6084)
          0.33333334 = coord(1/3)
      0.2 = coord(2/10)
    
    Abstract
    The challenge of cataloging Web sites and electronic resources was the most important issue facing the cataloging world in the last year. This article reviews attempts to analyze and revise the cataloging code in view of the new electronic environment. The difficulties of applying traditional library cataloging standards to Web resources has led some to favor metadata as the best means of providing access to these materials. The appropriate education and training for library cataloging personnel remains crucial during this transitional period. Articles on user understanding of Library of Congress subject headings and on cataloging practice are also reviewed.
    Date
    10. 9.2000 17:38:22
  4. Chambers, S.; Myall, C.: Cataloging and classification : review of the literature 2007-8 (2010) 0.01
    0.0065225097 = product of:
      0.032612547 = sum of:
        0.023567477 = weight(_text_:web in 4309) [ClassicSimilarity], result of:
          0.023567477 = score(doc=4309,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.25239927 = fieldWeight in 4309, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4309)
        0.009045068 = product of:
          0.027135205 = sum of:
            0.027135205 = weight(_text_:22 in 4309) [ClassicSimilarity], result of:
              0.027135205 = score(doc=4309,freq=2.0), product of:
                0.10019246 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.028611459 = queryNorm
                0.2708308 = fieldWeight in 4309, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4309)
          0.33333334 = coord(1/3)
      0.2 = coord(2/10)
    
    Abstract
    This paper surveys library literature on cataloging and classification published in 2007-8, indicating its extent and range in terms of types of literature, major subject areas, and themes. The paper reviews pertinent literature in the following areas: the future of bibliographic control, general cataloging standards and texts, Functional Requirements for Bibliographic Records (FRBR), cataloging varied resources, metadata and cataloging in the Web world, classification and subject access, questions of diversity and diverse perspectives, additional reports of practice and research, catalogers' education and careers, keeping current through columns and blogs, and cataloging history.
    Date
    10. 9.2000 17:38:22
  5. Chen, H.; Chau, M.: Web mining : machine learning for Web applications (2003) 0.01
    0.0063880207 = product of:
      0.063880205 = sum of:
        0.063880205 = weight(_text_:web in 4242) [ClassicSimilarity], result of:
          0.063880205 = score(doc=4242,freq=20.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.6841342 = fieldWeight in 4242, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4242)
      0.1 = coord(1/10)
    
    Abstract
    With more than two billion pages created by millions of Web page authors and organizations, the World Wide Web is a tremendously rich knowledge base. The knowledge comes not only from the content of the pages themselves, but also from the unique characteristics of the Web, such as its hyperlink structure and its diversity of content and languages. Analysis of these characteristics often reveals interesting patterns and new knowledge. Such knowledge can be used to improve users' efficiency and effectiveness in searching for information an the Web, and also for applications unrelated to the Web, such as support for decision making or business management. The Web's size and its unstructured and dynamic content, as well as its multilingual nature, make the extraction of useful knowledge a challenging research problem. Furthermore, the Web generates a large amount of data in other formats that contain valuable information. For example, Web server logs' information about user access patterns can be used for information personalization or improving Web page design.
  6. Large, A.: Children, teenagers, and the Web (2004) 0.01
    0.0053868517 = product of:
      0.053868517 = sum of:
        0.053868517 = weight(_text_:web in 4154) [ClassicSimilarity], result of:
          0.053868517 = score(doc=4154,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.5769126 = fieldWeight in 4154, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.125 = fieldNorm(doc=4154)
      0.1 = coord(1/10)
    
  7. Yang, K.: Information retrieval on the Web (2004) 0.01
    0.0053868517 = product of:
      0.053868517 = sum of:
        0.053868517 = weight(_text_:web in 4278) [ClassicSimilarity], result of:
          0.053868517 = score(doc=4278,freq=32.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.5769126 = fieldWeight in 4278, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=4278)
      0.1 = coord(1/10)
    
    Abstract
    How do we find information an the Web? Although information on the Web is distributed and decentralized, the Web can be viewed as a single, virtual document collection. In that regard, the fundamental questions and approaches of traditional information retrieval (IR) research (e.g., term weighting, query expansion) are likely to be relevant in Web document retrieval. Findings from traditional IR research, however, may not always be applicable in a Web setting. The Web document collection - massive in size and diverse in content, format, purpose, and quality - challenges the validity of previous research findings that are based an relatively small and homogeneous test collections. Moreover, some traditional IR approaches, although applicable in theory, may be impossible or impractical to implement in a Web setting. For instance, the size, distribution, and dynamic nature of Web information make it extremely difficult to construct a complete and up-to-date data representation of the kind required for a model IR system. To further complicate matters, information seeking on the Web is diverse in character and unpredictable in nature. Web searchers come from all walks of life and are motivated by many kinds of information needs. The wide range of experience, knowledge, motivation, and purpose means that searchers can express diverse types of information needs in a wide variety of ways with differing criteria for satisfying those needs. Conventional evaluation measures, such as precision and recall, may no longer be appropriate for Web IR, where a representative test collection is all but impossible to construct. Finding information on the Web creates many new challenges for, and exacerbates some old problems in, IR research. At the same time, the Web is rich in new types of information not present in most IR test collections. Hyperlinks, usage statistics, document markup tags, and collections of topic hierarchies such as Yahoo! (http://www.yahoo.com) present an opportunity to leverage Web-specific document characteristics in novel ways that go beyond the term-based retrieval framework of traditional IR. Consequently, researchers in Web IR have reexamined the findings from traditional IR research.
  8. Rasmussen, E.M.: Indexing and retrieval for the Web (2002) 0.00
    0.004409071 = product of:
      0.04409071 = sum of:
        0.04409071 = weight(_text_:web in 4285) [ClassicSimilarity], result of:
          0.04409071 = score(doc=4285,freq=28.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.47219574 = fieldWeight in 4285, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4285)
      0.1 = coord(1/10)
    
    Abstract
    The introduction and growth of the World Wide Web (WWW, or Web) have resulted in a profound change in the way individuals and organizations access information. In terms of volume, nature, and accessibility, the characteristics of electronic information are significantly different from those of even five or six years ago. Control of, and access to, this flood of information rely heavily an automated techniques for indexing and retrieval. According to Gudivada, Raghavan, Grosky, and Kasanagottu (1997, p. 58), "The ability to search and retrieve information from the Web efficiently and effectively is an enabling technology for realizing its full potential." Almost 93 percent of those surveyed consider the Web an "indispensable" Internet technology, second only to e-mail (Graphie, Visualization & Usability Center, 1998). Although there are other ways of locating information an the Web (browsing or following directory structures), 85 percent of users identify Web pages by means of a search engine (Graphie, Visualization & Usability Center, 1998). A more recent study conducted by the Stanford Institute for the Quantitative Study of Society confirms the finding that searching for information is second only to e-mail as an Internet activity (Nie & Ebring, 2000, online). In fact, Nie and Ebring conclude, "... the Internet today is a giant public library with a decidedly commercial tilt. The most widespread use of the Internet today is as an information search utility for products, travel, hobbies, and general information. Virtually all users interviewed responded that they engaged in one or more of these information gathering activities."
    Techniques for automated indexing and information retrieval (IR) have been developed, tested, and refined over the past 40 years, and are well documented (see, for example, Agosti & Smeaton, 1996; BaezaYates & Ribeiro-Neto, 1999a; Frakes & Baeza-Yates, 1992; Korfhage, 1997; Salton, 1989; Witten, Moffat, & Bell, 1999). With the introduction of the Web, and the capability to index and retrieve via search engines, these techniques have been extended to a new environment. They have been adopted, altered, and in some Gases extended to include new methods. "In short, search engines are indispensable for searching the Web, they employ a variety of relatively advanced IR techniques, and there are some peculiar aspects of search engines that make searching the Web different than more conventional information retrieval" (Gordon & Pathak, 1999, p. 145). The environment for information retrieval an the World Wide Web differs from that of "conventional" information retrieval in a number of fundamental ways. The collection is very large and changes continuously, with pages being added, deleted, and altered. Wide variability between the size, structure, focus, quality, and usefulness of documents makes Web documents much more heterogeneous than a typical electronic document collection. The wide variety of document types includes images, video, audio, and scripts, as well as many different document languages. Duplication of documents and sites is common. Documents are interconnected through networks of hyperlinks. Because of the size and dynamic nature of the Web, preprocessing all documents requires considerable resources and is often not feasible, certainly not an the frequent basis required to ensure currency. Query length is usually much shorter than in other environments-only a few words-and user behavior differs from that in other environments. These differences make the Web a novel environment for information retrieval (Baeza-Yates & Ribeiro-Neto, 1999b; Bharat & Henzinger, 1998; Huang, 2000).
  9. Legg, C.: Ontologies on the Semantic Web (2007) 0.00
    0.0038090795 = product of:
      0.038090795 = sum of:
        0.038090795 = weight(_text_:web in 1979) [ClassicSimilarity], result of:
          0.038090795 = score(doc=1979,freq=16.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.4079388 = fieldWeight in 1979, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=1979)
      0.1 = coord(1/10)
    
    Abstract
    As an informational technology, the World Wide Web has enjoyed spectacular success. In just ten years it has transformed the way information is produced, stored, and shared in arenas as diverse as shopping, family photo albums, and high-level academic research. The "Semantic Web" is touted by its developers as equally revolutionary, although it has not yet achieved anything like the Web's exponential uptake. It seeks to transcend a current limitation of the Web - that it largely requires indexing to be accomplished merely on specific character strings. Thus, a person searching for information about "turkey" (the bird) receives from current search engines many irrelevant pages about "Turkey" (the country) and nothing about the Spanish "pavo" even if he or she is a Spanish-speaker able to understand such pages. The Semantic Web vision is to develop technology to facilitate retrieval of information via meanings, not just spellings. For this to be possible, most commentators believe, Semantic Web applications will have to draw on some kind of shared, structured, machine-readable conceptual scheme. Thus, there has been a convergence between the Semantic Web research community and an older tradition with roots in classical Artificial Intelligence (AI) research (sometimes referred to as "knowledge representation") whose goal is to develop a formal ontology. A formal ontology is a machine-readable theory of the most fundamental concepts or "categories" required in order to understand information pertaining to any knowledge domain. A review of the attempts that have been made to realize this goal provides an opportunity to reflect in interestingly concrete ways on various research questions such as the following: - How explicit a machine-understandable theory of meaning is it possible or practical to construct? - How universal a machine-understandable theory of meaning is it possible or practical to construct? - How much (and what kind of) inference support is required to realize a machine-understandable theory of meaning? - What is it for a theory of meaning to be machine-understandable anyway?
    Theme
    Semantic Web
  10. Haythornthwaite, C.; Hagar, C.: ¬The social worlds of the Web (2004) 0.00
    0.0037641774 = product of:
      0.037641775 = sum of:
        0.037641775 = weight(_text_:web in 4282) [ClassicSimilarity], result of:
          0.037641775 = score(doc=4282,freq=10.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.40312994 = fieldWeight in 4282, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4282)
      0.1 = coord(1/10)
    
    Abstract
    We know this Web world. We live in it, particularly those of us in developed countries. Even if we do not go online daily, we live with itour culture is imprinted with online activity and vocabulary: e-mailing colleagues, surfing the Web, posting Web pages, blogging, gender-bending in cyberspace, texting and instant messaging friends, engaging in ecommerce, entering an online chat room, or morphing in an online world. We use it-to conduct business, find information, talk with friends and colleagues. We know it is something separate, yet we incorporate it into our daily lives. We identify with it, bringing to it behaviors and expectations we hold for the world in general. We approach it as explorers and entrepreneurs, ready to move into unknown opportunities and territory; creators and engineers, eager to build new structures; utopians for whom "the world of the Web" represents the chance to start again and "get it right" this time; utilitarians, ready to get what we can out of the new structures; and dystopians, for whom this is just more evidence that there is no way to "get it right." The word "world" has many connotations. The Oxford English Dictionary (http://dictionary.oed.com) gives 27 definitions for the noun "world" including: - The sphere within which one's interests are bound up or one's activities find scope; (one's) sphere of action or thought; the "realm" within which one moves or lives. - A group or system of things or beings associated by common characteristics (denoted by a qualifying word or phrase), or considered as constituting a unity. - Human society considered in relation to its activities, difficulties, temptations, and the like; hence, contextually, the ways, practices, or customs of the people among whom one lives; the occupations and interests of society at large.
  11. Marsh, S.; Dibben, M.R.: ¬The role of trust in information science and technology (2002) 0.00
    0.0034988632 = product of:
      0.03498863 = sum of:
        0.03498863 = weight(_text_:web in 4289) [ClassicSimilarity], result of:
          0.03498863 = score(doc=4289,freq=6.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.37471575 = fieldWeight in 4289, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4289)
      0.1 = coord(1/10)
    
    Abstract
    This chapter discusses the notion of trust as it relates to information science and technology, specifically user interfaces, autonomous agents, and information systems. We first present an in-depth discussion of the concept of trust in and of itself, moving an to applications and considerations of trust in relation to information technologies. We consider trust from a "soft" perspective-thus, although security concepts such as cryptography, virus protection, authentication, and so forth reinforce (or damage) the feelings of trust we may have in a system, they are not themselves constitutive of "trust." We discuss information technology from a human-centric viewpoint, where trust is a less well-structured but much more powerful phenomenon. With the proliferation of electronic commerce (e-commerce) and the World Wide Web (WWW, or Web), much has been made of the ability of individuals to explore the vast quantities of information available to them, to purchase goods (as diverse as vacations and cars) online, and to publish information an their personal Web sites.
  12. Denton, W.: Putting facets on the Web : an annotated bibliography (2003) 0.00
    0.0029157193 = product of:
      0.029157192 = sum of:
        0.029157192 = weight(_text_:web in 2467) [ClassicSimilarity], result of:
          0.029157192 = score(doc=2467,freq=24.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.3122631 = fieldWeight in 2467, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.01953125 = fieldNorm(doc=2467)
      0.1 = coord(1/10)
    
    Abstract
    This is a classified, annotated bibliography about how to design faceted classification systems and make them usable on the World Wide Web. It is the first of three works I will be doing. The second, based on the material here and elsewhere, will discuss how to actually make the faceted system and put it online. The third will be a report of how I did just that, what worked, what didn't, and what I learned. Almost every article or book listed here begins with an explanation of what a faceted classification system is, so I won't (but see Steckel in Background below if you don't already know). They all agree that faceted systems are very appropriate for the web. Even pre-web articles (such as Duncan's in Background, below) assert that hypertext and facets will go together well. Combined, it is possible to take a set of documents and classify them or apply subject headings to describe what they are about, then build a navigational structure so that any user, no matter how he or she approaches the material, no matter what his or her goals, can move and search in a way that makes sense to them, but still get to the same useful results as someone else following a different path to the same goal. There is no one way that everyone will always use when looking for information. The more flexible the organization of the information, the more accommodating it is. Facets are more flexible for hypertext browsing than any enumerative or hierarchical system.
    Consider movie listings in newspapers. Most Canadian newspapers list movie showtimes in two large blocks, for the two major theatre chains. The listings are ordered by region (in large cities), then theatre, then movie, and finally by showtime. Anyone wondering where and when a particular movie is playing must scan the complete listings. Determining what movies are playing in the next half hour is very difficult. When movie listings went onto the web, most sites used a simple faceted organization, always with movie name and theatre, and perhaps with region or neighbourhood (thankfully, theatre chains were left out). They make it easy to pick a theatre and see what movies are playing there, or to pick a movie and see what theatres are showing it. To complete the system, the sites should allow users to browse by neighbourhood and showtime, and to order the results in any way they desired. Thus could people easily find answers to such questions as, "Where is the new James Bond movie playing?" "What's showing at the Roxy tonight?" "I'm going to be out in in Little Finland this afternoon with three hours to kill starting at 2 ... is anything interesting playing?" A hypertext, faceted classification system makes more useful information more easily available to the user. Reading the books and articles below in chronological order will show a certain progression: suggestions that faceting and hypertext might work well, confidence that facets would work well if only someone would make such a system, and finally the beginning of serious work on actually designing, building, and testing faceted web sites. There is a solid basis of how to make faceted classifications (see Vickery in Recommended), but their application online is just starting. Work on XFML (see Van Dijck's work in Recommended) the Exchangeable Faceted Metadata Language, will make this easier. If it follows previous patterns, parts of the Internet community will embrace the idea and make open source software available for others to reuse. It will be particularly beneficial if professionals in both information studies and computer science can work together to build working systems, standards, and code. Each can benefit from the other's expertise in what can be a very complicated and technical area. One particularly nice thing about this area of research is that people interested in combining facets and the web often have web sites where they post their writings.
    This bibliography is not meant to be exhaustive, but unfortunately it is not as complete as I wanted. Some books and articles are not be included, but they may be used in my future work. (These include two books and one article by B.C. Vickery: Faceted Classification Schemes (New Brunswick, NJ: Rutgers, 1966), Classification and Indexing in Science, 3rd ed. (London: Butterworths, 1975), and "Knowledge Representation: A Brief Review" (Journal of Documentation 42 no. 3 (September 1986): 145-159; and A.C. Foskett's "The Future of Faceted Classification" in The Future of Classification, edited by Rita Marcella and Arthur Maltby (Aldershot, England: Gower, 2000): 69-80). Nevertheless, I hope this bibliography will be useful for those both new to or familiar with faceted hypertext systems. Some very basic resources are listed, as well as some very advanced ones. Some example web sites are mentioned, but there is no detailed technical discussion of any software. The user interface to any web site is extremely important, and this is briefly mentioned in two or three places (for example the discussion of lawforwa.org (see Example Web Sites)). The larger question of how to display information graphically and with hypertext is outside the scope of this bibliography. There are five sections: Recommended, Background, Not Relevant, Example Web Sites, and Mailing Lists. Background material is either introductory, advanced, or of peripheral interest, and can be read after the Recommended resources if the reader wants to know more. The Not Relevant category contains articles that may appear in bibliographies but are not relevant for my purposes.
  13. Yu, N.: Readings & Web resources for faceted classification 0.00
    0.0028568096 = product of:
      0.028568096 = sum of:
        0.028568096 = weight(_text_:web in 4394) [ClassicSimilarity], result of:
          0.028568096 = score(doc=4394,freq=4.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.3059541 = fieldWeight in 4394, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4394)
      0.1 = coord(1/10)
    
    Abstract
    The term "facet" has been used in various places, while in most cases it is just a buzz word to replace what is indeed "aspect" or "category". The references below either define and explain the original concept of facet or provide guidelines for building 'real' faceted search/browse. I was interested in faceted classification because it seems to be a natural and efficient way for organizing and browsing Web collections. However, to automatically generate facets and their isolates is extremely difficult since it involves concept extraction and concept grouping, both of which are difficult problems by themselves. And it is almost impossible to achieve mutually exclusive and jointly exhaustive 'true' facets without human judgment. Nowadays, faceted search/browse widely exists, implicitly or explicitly, on a majority of retail websites due to the multi-aspects nature of the data. However, it is still rarely seen on any digital library sites. (I could be wrong since I haven't kept myself updated with this field for a while.)
  14. El-Sherbini, M.: Selected cataloging tools on the Internet (2003) 0.00
    0.0026934259 = product of:
      0.026934259 = sum of:
        0.026934259 = weight(_text_:web in 1997) [ClassicSimilarity], result of:
          0.026934259 = score(doc=1997,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.2884563 = fieldWeight in 1997, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0625 = fieldNorm(doc=1997)
      0.1 = coord(1/10)
    
    Abstract
    This bibliography contains selected cataloging tools an the Internet. It is divided into seven sections as follows: authority management and subject headings tools; cataloging tools by type of materials; dictionaries, encyclopedias, and place names; listservs and workshops; software and vendors; technical service professional organizations; and journals and newsletters. Resources are arranged in alphabetical order under each topic. Selected cataloging tools are annotated. There is some overlap since a given web site can cover many tools.
  15. Høyrup, E.: Books about mathematics : history, philosophy, education, models, system theory, and works of reference etc (1979) 0.00
    0.0025815545 = product of:
      0.025815545 = sum of:
        0.025815545 = product of:
          0.07744663 = sum of:
            0.07744663 = weight(_text_:29 in 1733) [ClassicSimilarity], result of:
              0.07744663 = score(doc=1733,freq=4.0), product of:
                0.10064617 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.028611459 = queryNorm
                0.7694941 = fieldWeight in 1733, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1733)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Date
    9. 1.1999 18:29:29
  16. Siqueira, J.; Martins, D.L.: Workflow models for aggregating cultural heritage data on the web : a systematic literature review (2022) 0.00
    0.0023806747 = product of:
      0.023806747 = sum of:
        0.023806747 = weight(_text_:web in 464) [ClassicSimilarity], result of:
          0.023806747 = score(doc=464,freq=4.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.25496176 = fieldWeight in 464, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=464)
      0.1 = coord(1/10)
    
    Abstract
    In recent years, different cultural institutions have made efforts to spread culture through the construction of a unique search interface that integrates their digital objects and facilitates data retrieval for lay users. However, integrating cultural data is not a trivial task; therefore, this work performs a systematic literature review on data aggregation workflows, in order to answer five questions: What are the projects? What are the planned steps? Which technologies are used? Are the steps performed manually, automatically, or semi-automatically? Which perform semantic search? The searches were carried out in three databases: Networked Digital Library of Theses and Dissertations, Scopus and Web of Science. In Q01, 12 projects were selected. In Q02, 9 stages were identified: Harvesting, Ingestion, Mapping, Indexing, Storing, Monitoring, Enriching, Displaying, and Publishing LOD. In Q03, 19 different technologies were found it. In Q04, we identified that most of the solutions are semi-automatic and, in Q05, that most of them perform a semantic search. The analysis of the workflows allowed us to identify that there is no consensus regarding the stages, their nomenclatures, and technologies, besides presenting superficial discussions. But it allowed to identify the main steps for the implementation of the aggregation of cultural data.
  17. Martin, B.: Knowledge management (2008) 0.00
    0.002086211 = product of:
      0.02086211 = sum of:
        0.02086211 = product of:
          0.06258633 = sum of:
            0.06258633 = weight(_text_:29 in 4230) [ClassicSimilarity], result of:
              0.06258633 = score(doc=4230,freq=2.0), product of:
                0.10064617 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.028611459 = queryNorm
                0.6218451 = fieldWeight in 4230, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.125 = fieldNorm(doc=4230)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Date
    13. 7.2008 9:29:38
  18. Priss, U.: Formal concept analysis in information science (2006) 0.00
    0.002086211 = product of:
      0.02086211 = sum of:
        0.02086211 = product of:
          0.06258633 = sum of:
            0.06258633 = weight(_text_:29 in 4305) [ClassicSimilarity], result of:
              0.06258633 = score(doc=4305,freq=2.0), product of:
                0.10064617 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.028611459 = queryNorm
                0.6218451 = fieldWeight in 4305, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.125 = fieldNorm(doc=4305)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Date
    13. 7.2008 19:29:59
  19. Enser, P.G.B.: Visual image retrieval (2008) 0.00
    0.0020674444 = product of:
      0.020674443 = sum of:
        0.020674443 = product of:
          0.062023327 = sum of:
            0.062023327 = weight(_text_:22 in 3281) [ClassicSimilarity], result of:
              0.062023327 = score(doc=3281,freq=2.0), product of:
                0.10019246 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.028611459 = queryNorm
                0.61904186 = fieldWeight in 3281, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=3281)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Date
    22. 1.2012 13:01:26
  20. Morris, S.A.: Mapping research specialties (2008) 0.00
    0.0020674444 = product of:
      0.020674443 = sum of:
        0.020674443 = product of:
          0.062023327 = sum of:
            0.062023327 = weight(_text_:22 in 3962) [ClassicSimilarity], result of:
              0.062023327 = score(doc=3962,freq=2.0), product of:
                0.10019246 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.028611459 = queryNorm
                0.61904186 = fieldWeight in 3962, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=3962)
          0.33333334 = coord(1/3)
      0.1 = coord(1/10)
    
    Date
    13. 7.2008 9:30:22

Languages

  • e 64
  • d 1
  • m 1
  • ru 1
  • More… Less…

Types

  • a 60
  • b 15
  • m 3
  • el 2
  • r 1
  • s 1
  • More… Less…