Search (215 results, page 11 of 11)

  • × type_ss:"m"
  • × year_i:[2000 TO 2010}
  1. Willinsky, J.: ¬The access principle : the case for open access to research and scholarship (2006) 0.00
    0.0033514309 = product of:
      0.010054292 = sum of:
        0.010054292 = product of:
          0.020108584 = sum of:
            0.020108584 = weight(_text_:indexing in 298) [ClassicSimilarity], result of:
              0.020108584 = score(doc=298,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.105730906 = fieldWeight in 298, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=298)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Footnote
    Rez. in: JASIST 58(2007) no.9, S.1386 (L.A. Ennis): "Written by John Willinsky. Pacific Press Professor of Literacy and Technology at the University of British Columbia and Open Journals Systems Software des eloper. the eighth hook in the Digital Libraries and Electronic Publishing series (edited by William Y. Arms) provides a compelling and convincing argument in favor of open access. At the core of this work is Willinsky's "access principle." a commitment that "research carries with it a responsibility to extend circulation of such work as far as possible and ideally to all who are interested in it and all who might profit from it" (p.xii). One by one Willinsky tackles the obstacles. both real and perceived, to open access. succeeding in his goal to "inform and inspire a larger debate over the political and moral economy of knowledge" (p.xiv). The author does note the irony of publishing a hook while advocating for open access, but points out that he does so to reach a larger audience. Willinsky also points out that most of the chapters' earlier versions can be found in open-access journals and on his Web site (http://www.11ed.educubc.ca/faculty/willinsky.html). The Access Principle is organized topically into thirteen chapters covering a broad range of practical and theoretical issues. Taken together. these chapters provide the reader with an excellent introduction to the open-access debate as well as all the potential benefits and possible impacts of the open-access movement. The author also includes six appendices. with information on metadata and indexing. os er twenty pages of references, and an index. ... All of Willinsky's arguments arc convincing and heartfelt. It is apparent throughout the hook that the author deeply believes in the principles behind open access. and his passion and conviction come through in the work. making the hook a thought-provoking and very interesting read. While he offers numerous examples to illustrate his points throughout the work. he does not. however. offer solutions or state that he has all the answers. In that, he succeeds in his goal to craft a hook that "informs and inspires. As a result, The Access Principle is an important read for information professionals, researchers, and academics of all kinds, whether or not the reader agrees with Willinsky."
  2. Kochtanek, T.R.; Matthews, J.R.: Library information systems : from library automation to distributed information systems (2002) 0.00
    0.0033514309 = product of:
      0.010054292 = sum of:
        0.010054292 = product of:
          0.020108584 = sum of:
            0.020108584 = weight(_text_:indexing in 1792) [ClassicSimilarity], result of:
              0.020108584 = score(doc=1792,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.105730906 = fieldWeight in 1792, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=1792)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Footnote
    Rez. in: JASIST 54(2003) no.12, S.1166-1167 (Brenda Chawner): "Kochtanek and Matthews have written a welcome addition to the small set of introductory texts an applications of information technology to library and information Services. The book has fourteen chapters grouped into four sections: "The Broader Context," "The Technologies," "Management Issues," and "Future Considerations." Two chapters provide the broad content, with the first giving a historical overview of the development and adoption of "library information systems." Kochtanek and Matthews define this as "a wide array of solutions that previously might have been considered separate industries with distinctly different marketplaces" (p. 3), referring specifically to integrated library systems (ILS, and offen called library management systems in this part of the world), and online databases, plus the more recent developments of Web-based resources, digital libraries, ebooks, and ejournals. They characterize technology adoption patterns in libraries as ranging from "bleeding edge" to "leading edge" to "in the wedge" to "trailing edge"-this is a catchy restatement of adopter categories from Rogers' diffusion of innovation theory, where they are more conventionally known as "early adopters," "early majority," "late majority," and "laggards." This chapter concludes with a look at more general technology trends that have affected library applications, including developments in hardware (moving from mainframes to minicomputers to personal Computers), changes in software development (from in-house to packages), and developments in communications technology (from dedicated host Computers to more open networks to the current distributed environment found with the Internet). This is followed by a chapter describing the ILS and online database industries in some detail. "The Technologies" begins with a chapter an the structure and functionality of integrated library systems, which also includes a brief discussion of precision versus recall, managing access to internal documents, indexing and searching, and catalogue maintenance. This is followed by a chapter an open systems, which concludes with a useful list of questions to consider to determine an organization's readiness to adopt open source solutions. As one world expect, this section also includes a detailed chapter an telecommunications and networking, which includes types of networks, transmission media, network topologies, switching techniques (ranging from dial up and leased lines to ISDN/DSL, frame relay, and ATM). It concludes with a chapter an the role and importance of standards, which covers the need for standards and standards organizations, and gives examples of different types of standards, such as MARC, Dublin Core, Z39.50, and markup standards such as SGML, HTML, and XML. Unicode is also covered but only briefly. This section world be strengthened by a chapter an hardware concepts-the authors assume that their reader is already familiar with these, which may not be true in all cases (for example, the phrase "client-Server" is first used an page 11, but only given a brief definition in the glossary). Burke's Library Technology Companion: A Basic Guide for Library Staff (New York: Neal-Schuman, 2001) might be useful to fill this gap at an introductory level, and Saffady's Introduction to Automation for Librarians, 4th ed. (Chicago: American Library Association, 1999) world be better for those interested in more detail. The final two sections, however, are the book's real strength, with a strong focus an management issues, and this content distinguishes it from other books an this topic such as Ferguson and Hebels Computers for Librarians: an Introduction to Systems and Applications (Waggawagga, NSW: Centre for Information Studies, Charles Sturt University, 1998). ...
  3. Emerging frameworks and methods : Proceedings of the Fourth International Conference on the Conceptions of Library and Information Science (CoLIS4), Seattle, WA, July 21 - 25, 2002 (2002) 0.00
    0.0031733024 = product of:
      0.009519907 = sum of:
        0.009519907 = product of:
          0.019039813 = sum of:
            0.019039813 = weight(_text_:22 in 55) [ClassicSimilarity], result of:
              0.019039813 = score(doc=55,freq=4.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.109432176 = fieldWeight in 55, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.015625 = fieldNorm(doc=55)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 2.2007 18:56:23
    22. 2.2007 19:12:10
  4. Devlin, K.: ¬Der Mathe-Instinkt (2005) 0.00
    0.0028048295 = product of:
      0.008414488 = sum of:
        0.008414488 = product of:
          0.016828977 = sum of:
            0.016828977 = weight(_text_:22 in 5944) [ClassicSimilarity], result of:
              0.016828977 = score(doc=5944,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.09672529 = fieldWeight in 5944, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=5944)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 7.2006 20:12:41
  5. ¬Die Zukunft des Wissens : Vorträge und Kolloquien: XVIII. Deutscher Kongress für Philosophie, Konstanz, 4. - 8. Oktober 1999 (2000) 0.00
    0.0028048295 = product of:
      0.008414488 = sum of:
        0.008414488 = product of:
          0.016828977 = sum of:
            0.016828977 = weight(_text_:22 in 733) [ClassicSimilarity], result of:
              0.016828977 = score(doc=733,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.09672529 = fieldWeight in 733, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=733)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 6.2005 15:30:21
  6. Lipow, A.G.: ¬The virtual reference librarian's handbook (2003) 0.00
    0.0028048295 = product of:
      0.008414488 = sum of:
        0.008414488 = product of:
          0.016828977 = sum of:
            0.016828977 = weight(_text_:22 in 3992) [ClassicSimilarity], result of:
              0.016828977 = score(doc=3992,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.09672529 = fieldWeight in 3992, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=3992)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 3.2004 14:46:50
  7. Umstätter, W.; Wagner-Döbler, R.; Löffler, K.: Einführung in die Katalogkunde : Vom Zettelkatalog zur Suchmaschine (2005) 0.00
    0.0028048295 = product of:
      0.008414488 = sum of:
        0.008414488 = product of:
          0.016828977 = sum of:
            0.016828977 = weight(_text_:22 in 4880) [ClassicSimilarity], result of:
              0.016828977 = score(doc=4880,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.09672529 = fieldWeight in 4880, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=4880)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 1.2006 19:45:49
  8. Strzolka, R.: ¬Das Internet als Weltbibliothek : Suchmaschinen und ihre Bedeutung für den Wissenserwerb (2008) 0.00
    0.0028048295 = product of:
      0.008414488 = sum of:
        0.008414488 = product of:
          0.016828977 = sum of:
            0.016828977 = weight(_text_:22 in 2331) [ClassicSimilarity], result of:
              0.016828977 = score(doc=2331,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.09672529 = fieldWeight in 2331, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=2331)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Footnote
    Neben Anwendungen und Beispielen aus verschiedenen Ländern belegt der Autor anhand beeindruckender Beispiele, wie wenig sinnstiftend Wort-/Wortsstammfilter sind (Seite 84 bis 91). Ein Hauptproblem in der Informations- und Wissensgesellschaft sieht er darin: »Die Bequemlichkeit der Nutzer« führe zu einer selbst auferlegten »Informationsbeschränkung« bei der Nutzung von Suchmaschinen (Seite 18 bis 21). Dementsprechend sieht Strzolka in der »fehlenden Informationskompetenz und Zensur« eine »unheilvolle Allianz« (Seite 24). »Sich auf Suchmaschinen als objektive Instanz zu verlassen«, heißt es später (Seite 56), »ist so naiv, wie die Bibel als einzige Informationsquelle für das Leben in der modernen Welt zu nutzen«. Denn: »Suchmaschinen zensieren nicht - Ihre Betreiber organisieren Zensur.« Internetzensur Auf 25 Seiten (Seite 74 bis 99) dokumentiert Strzolka mit bemerkenswerten Beispielen, wie das in der Praxis geschieht. Nach Strzolka beteiligen sich Bibliothekare auch an der (Internet-)Zensur (Seite 22 f, mit Beispielen auf Seite 34). Hier wäre eine tiefergehende Auseinsandersetzung angebracht gewesen. Beeindruckend dabei, wie Strzolka über »Zensur und Idiotenfallen« (Seite 84 bis 90) berichtet. »Das Internet ist eine ausgesprochen heterogene Datensammlung und >keineswegs die größte Bibliothek der Welt>, wie immer zu hören ist.« (Seite 25) Denn: »Informationsportale verengen den Weg in die Wissensgesellschaft [...], weil sie die Fokussierung auf gewünschte Quellen [fördern]« (Seite 25). Vor diesem Hintergrund müsse deshalb eher von »Informationsbeschränkung« als von »Informationsvielfalt« die Rede sein.
  9. Information science in transition (2009) 0.00
    0.0028048295 = product of:
      0.008414488 = sum of:
        0.008414488 = product of:
          0.016828977 = sum of:
            0.016828977 = weight(_text_:22 in 634) [ClassicSimilarity], result of:
              0.016828977 = score(doc=634,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.09672529 = fieldWeight in 634, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=634)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 2.2013 11:35:35
  10. XML data management : native XML and XML-enabled database systems (2003) 0.00
    0.002681145 = product of:
      0.0080434345 = sum of:
        0.0080434345 = product of:
          0.016086869 = sum of:
            0.016086869 = weight(_text_:indexing in 2073) [ClassicSimilarity], result of:
              0.016086869 = score(doc=2073,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.08458473 = fieldWeight in 2073, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.015625 = fieldNorm(doc=2073)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Footnote
    After several detailed examples of XML, Direen and Jones discuss sequence comparisons. The ability to create scored comparisons by such techniques as sequence alignment is fundamental to bioinformatics. For example, the function of a gene product may be inferred from similarity with a gene of known function but originating from a different organism and any information modeling method must facilitate such comparisons. One such comparison tool, BLAST utilizes a heuristic method has become the tool of choice for many years and is integrated into the NeoCore XMS (XML Management System) described herein. Any set of sequences that can be identified using an XPath query may thus become the targets of an embedded search. Again examples are given, though a BLASTp (protein) search is labeled as being BLASTn (nucleotide sequence) in one of them. Some variants of BLAST are computationally intensive, e.g., tBLASTx where a nucleotide sequence is dynamically translated in all six reading frames and compared against similarly translated database sequences. Though these variants are implemented in NeoCore XMS, it would be interesting to see runtimes for such comparisons. Obviously the utility of this and the other four quite specific examples will depend an your interest in the application area but two that are more research-oriented and general follow them. These chapters (on using XML with inductive databases and an XML warehouses) are both readable critical reviews of their respective subject areas. For those involved in the implementation of performance-critical applications an examination of benchmark results is mandatory, however very few would examine the benchmark tests themselves. The picture that emerges from this section is that no single set is comprehensive and that some functionalities are not addressed by any available benchmark. As always, there is no Substitute for an intimate knowledge of your data and how it is used. In a direct comparison of an XML-enabled and a native XML database system (unfortunately neither is named), the authors conclude that though the native system has the edge in handling large documents this comes at the expense of increasing index and data file size. The need to use legacy data and software will certainly favor the all-pervasive XML-enabled RDBMS such as Oracle 9i and IBM's DB2. Of more general utility is the chapter by Schmauch and Fellhauer comparing the approaches used by database systems for the storing of XML documents. Many of the limitations of current XML-handling systems may be traced to problems caused by the semi-structured nature of the documents and while the authors have no panacea, the chapter forms a useful discussion of the issues and even raises the ugly prospect that a return to the drawing board may be unavoidable. The book concludes with an appraisal of the current status of XML by the editors that perhaps focuses a little too little an the database side but overall I believe this book to be very useful indeed. Some of the indexing is a little idiosyncratic, for example some tags used in the examples are indexed (perhaps a separate examples index would be better) and Ron Bourret's excellent web site might be better placed under "Bourret" rather than under "Ron" but this doesn't really detract from the book's qualities. The broad spectrum and careful balance of theory and practice is a combination that both database and XML professionals will find valuable."
  11. Mandl, T.: Tolerantes Information Retrieval : Neuronale Netze zur Erhöhung der Adaptivität und Flexibilität bei der Informationssuche (2001) 0.00
    0.002681145 = product of:
      0.0080434345 = sum of:
        0.0080434345 = product of:
          0.016086869 = sum of:
            0.016086869 = weight(_text_:indexing in 5965) [ClassicSimilarity], result of:
              0.016086869 = score(doc=5965,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.08458473 = fieldWeight in 5965, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.015625 = fieldNorm(doc=5965)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Footnote
    da nun nach fast 200 Seiten der Hauptteil der Dissertation folgt - die Vorstellung und Bewertung des bereits erwähnten COSIMIR Modells. Das COSIMIR Modell "berechnet die Ähnlichkeit zwischen den zwei anliegenden Input-Vektoren" (P.194). Der Output des Netzwerks wird an einem einzigen Knoten abgegriffen, an dem sich ein sogenannten Relevanzwert einstellt, wenn die Berechnungen der Gewichtungen interner Knoten zum Abschluss kommen. Diese Gewichtungen hängen von den angelegten Inputvektoren, aus denen die Gewichte der ersten Knotenschicht ermittelt werden, und den im Netzwerk vorgegebenen Kantengewichten ab. Die Gewichtung von Kanten ist der Kernpunkt des neuronalen Ansatzes: In Analogie zum biologischen Urbild (Dendrit mit Synapsen) wächst das Gewicht der Kante mit jeder Aktivierung während einer Trainingsphase. Legt man in dieser Phase zwei Inputvektoren, z.B. Dokumentvektor und Ouery gleichzeitig mit dem Relevanzurteil als Wert des Outputknoten an, verteilen sich durch den BackpropagationProzess die Gewichte entlang der Pfade, die zwischen den beteiligten Knoten bestehen. Da alle Knoten miteinander verbunden sind, entstehen nach mehreren Trainingsbeispielen bereits deutlich unterschiedliche Kantengewichte, weil die aktiv beteiligten Kanten die Änderungen akkumulativ speichern. Eine Variation des Verfahrens benutzt das NN als "Transformationsnetzwerk", wobei die beiden Inputvektoren mit einer Dokumentrepräsentation und einem dazugehörigen Indexat (von einem Experten bereitgestellt) belegt werden. Neben der schon aufgezeigten Trainingsnotwendigkeit weisen die Neuronalen Netze eine weitere intrinsische Problematik auf: Je mehr äußere Knoten benötigt werden, desto mehr interne Kanten (und bei der Verwendung von Zwischenschichten auch Knoten) sind zu verwalten, deren Anzahl nicht linear wächst. Dieser algorithmische Befund setzt naiven Einsätzen der NN-Modelle in der Praxis schnell Grenzen, deshalb ist es umso verdienstvoller, dass der Autor einen innovativen Weg zur Lösung des Problems mit den Mitteln des IR vorschlagen kann. Er verwendet das Latent Semantic Indexing, welches Dokumentrepräsentationen aus einem hochdimensionalen Vektorraum in einen niederdimensionalen abbildet, um die Anzahl der Knoten deutlich zu reduzieren. Damit ist eine sehr schöne Synthese gelungen, welche die eingangs angedeuteten formalen Übereinstimmungen zwischen Vektorraummodellen im IR und den NN aufzeigt und ausnutzt.
  12. Lewandowski, D.: Web Information Retrieval : Technologien zur Informationssuche im Internet (2005) 0.00
    0.002681145 = product of:
      0.0080434345 = sum of:
        0.0080434345 = product of:
          0.016086869 = sum of:
            0.016086869 = weight(_text_:indexing in 3635) [ClassicSimilarity], result of:
              0.016086869 = score(doc=3635,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.08458473 = fieldWeight in 3635, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.015625 = fieldNorm(doc=3635)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Content
    7 Informationsstatistische und informationslinguistische Verfahren 7.1 Textstatistische Verfahren 7.2 Nutzungsstatistische Verfahren 7.3 Informationslinguistische Verfahren 8 Linktopologische Rankingverfahren 8.1 Grundlagen linktopologischer Rankingverfahren: Science Citation Indexing 8.2 PageRank 8.3 Kleinbergs HITS 8.4 Hilltop 8.5 Evaluierung linktopologischer Verfahren 8.6 Problembereiche linktopologischer Rankingverfahren 8.7 Fazit linktopologische Verfahren 9 Retrievaltests 9.1 Aufbau und Nutzen von Retrievaltests 9.2 Aufbau und Ergebnisse ausgewählter Retrievaltests 9.3 Kritik an Retrievaltests 10 Verfahren der intuitiven Benutzerführung 10.1 Relevance Feedback 10.2 Vorschläge zur Erweiterung und Einschränkung der Suchanfrage 10.3 Klassifikation und Thesaurus 10.4 Clusterbildung 11 Aktualität 11.1 Bedeutung der Beschränkung nach der Aktualität der Dokumente 11.2 Funktionsfähigkeit der Datumsbeschränkung in Suchmaschinen 11.3 Möglichkeiten der Ermittlung von Datumsangaben in Web-Dokumenten 11.4 Aktualitätsfaktoren im Ranking 11.5 Spezialisierte Suchmaschinen für Nachrichten 11.6 Auswahl der gewünschten Aktualität durch den Nutzer 12 Qualität 12.1 Bedeutung der Beschränkung nach der Qualität der Dokumente 12.2 Qualitätsbeschränkungen bei der Recherche in Datenbank-Hosts 12.3 Identifizierung von Top-Quellen im WWW 12.4 Manuelle Einbindung von Top-Quellen 12.5 Automatisierte Einbindung von Invisible-Web-Quellen 12.6 Einbindung von Web-Verzeichnissen in Suchmaschinen 13 Verbesserung der Dokumentrepräsentation 13.1 Beschränkung auf den Inhaltsteil der Dokumente 13.2 Erweiterungen der Dokumentrepräsentation 13.3 Ersatz für die Nicht-Verwendbarkeit generischer Top-Level-Domains 13.4 Aufbereitung der Suchergebnisse in den Trefferlisten 14 Fazit und Ausblick 15 Literaturverzeichnis
  13. Bade, D.: ¬The creation and persistence of misinformation in shared library catalogs : language and subject knowledge in a technological era (2002) 0.00
    0.0022438637 = product of:
      0.0067315907 = sum of:
        0.0067315907 = product of:
          0.013463181 = sum of:
            0.013463181 = weight(_text_:22 in 1858) [ClassicSimilarity], result of:
              0.013463181 = score(doc=1858,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.07738023 = fieldWeight in 1858, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.015625 = fieldNorm(doc=1858)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 9.1997 19:16:05
  14. Information visualization in data mining and knowledge discovery (2002) 0.00
    0.0022438637 = product of:
      0.0067315907 = sum of:
        0.0067315907 = product of:
          0.013463181 = sum of:
            0.013463181 = weight(_text_:22 in 1789) [ClassicSimilarity], result of:
              0.013463181 = score(doc=1789,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.07738023 = fieldWeight in 1789, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.015625 = fieldNorm(doc=1789)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    23. 3.2008 19:10:22
  15. Bruce, H.: ¬The user's view of the Internet (2002) 0.00
    0.0016828976 = product of:
      0.0050486927 = sum of:
        0.0050486927 = product of:
          0.010097385 = sum of:
            0.010097385 = weight(_text_:22 in 4344) [ClassicSimilarity], result of:
              0.010097385 = score(doc=4344,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.058035173 = fieldWeight in 4344, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.01171875 = fieldNorm(doc=4344)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Footnote
    Chapter 2 (Technology and People) focuses an several theories of technological acceptance and diffusion. Unfortunately, Bruce's presentation is somewhat confusing as he moves from one theory to next, never quite connecting them into a logical sequence or coherent whole. Two theories are of particular interest to Bruce: the Theory of Diffusion of Innovations and the Theory of Planned Behavior. The Theory of Diffusion of Innovations is an "information-centric view of technology acceptance" in which technology adopters are placed in the information flows of society from which they learn about innovations and "drive innovation adoption decisions" (p. 20). The Theory of Planned Behavior maintains that the "performance of a behavior is a joint function of intentions and perceived behavioral control" (i.e., how muck control a person thinks they have) (pp. 22-23). Bruce combines these two theories to form the basis for the Technology Acceptance Model. This model posits that "an individual's acceptance of information technology is based an beliefs, attitudes, intentions, and behaviors" (p. 24). In all these theories and models echoes a recurring theme: "individual perceptions of the innovation or technology are critical" in terms of both its characteristics and its use (pp. 24-25). From these, in turn, Bruce derives a predictive theory of the role personal perceptions play in technology adoption: Personal Innovativeness of Information Technology Adoption (PIITA). Personal inventiveness is defined as "the willingness of an individual to try out any new information technology" (p. 26). In general, the PIITA theory predicts that information technology will be adopted by individuals that have a greater exposure to mass media, rely less an the evaluation of information technology by others, exhibit a greater ability to cope with uncertainty and take risks, and requires a less positive perception of an information technology prior to its adoption. Chapter 3 (A Focus an Usings) introduces the User-Centered Paradigm (UCP). The UCP is characteristic of the shift of emphasis from technology to users as the driving force behind technology and research agendas for Internet development [for a dissenting view, see Andrew Dillion's (2003) challenge to the utility of user-centerness for design guidance]. It entails the "broad acceptance of the user-oriented perspective across a range of disciplines and professional fields," such as business, education, cognitive engineering, and information science (p. 34).

Authors

Languages

Types

  • s 70
  • i 4
  • b 1
  • el 1
  • n 1
  • More… Less…

Subjects

Classifications