Search (45 results, page 2 of 3)

  • × classification_ss:"06.74 / Informationssysteme"
  • × language_ss:"e"
  • × type_ss:"m"
  1. TREC: experiment and evaluation in information retrieval (2005) 0.01
    0.005513061 = product of:
      0.013782652 = sum of:
        0.0061394903 = weight(_text_:a in 636) [ClassicSimilarity], result of:
          0.0061394903 = score(doc=636,freq=26.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.11483221 = fieldWeight in 636, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.01953125 = fieldNorm(doc=636)
        0.0076431613 = product of:
          0.015286323 = sum of:
            0.015286323 = weight(_text_:information in 636) [ClassicSimilarity], result of:
              0.015286323 = score(doc=636,freq=30.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.18779588 = fieldWeight in 636, product of:
                  5.477226 = tf(freq=30.0), with freq of:
                    30.0 = termFreq=30.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=636)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    The Text REtrieval Conference (TREC), a yearly workshop hosted by the US government's National Institute of Standards and Technology, provides the infrastructure necessary for large-scale evaluation of text retrieval methodologies. With the goal of accelerating research in this area, TREC created the first large test collections of full-text documents and standardized retrieval evaluation. The impact has been significant; since TREC's beginning in 1992, retrieval effectiveness has approximately doubled. TREC has built a variety of large test collections, including collections for such specialized retrieval tasks as cross-language retrieval and retrieval of speech. Moreover, TREC has accelerated the transfer of research ideas into commercial systems, as demonstrated in the number of retrieval techniques developed in TREC that are now used in Web search engines. This book provides a comprehensive review of TREC research, summarizing the variety of TREC results, documenting the best practices in experimental information retrieval, and suggesting areas for further research. The first part of the book describes TREC's history, test collections, and retrieval methodology. Next, the book provides "track" reports -- describing the evaluations of specific tasks, including routing and filtering, interactive retrieval, and retrieving noisy text. The final part of the book offers perspectives on TREC from such participants as Microsoft Research, University of Massachusetts, Cornell University, University of Waterloo, City University of New York, and IBM. The book will be of interest to researchers in information retrieval and related technologies, including natural language processing.
    Content
    Enthält die Beiträge: 1. The Text REtrieval Conference - Ellen M. Voorhees and Donna K. Harman 2. The TREC Test Collections - Donna K. Harman 3. Retrieval System Evaluation - Chris Buckley and Ellen M. Voorhees 4. The TREC Ad Hoc Experiments - Donna K. Harman 5. Routing and Filtering - Stephen Robertson and Jamie Callan 6. The TREC Interactive Tracks: Putting the User into Search - Susan T. Dumais and Nicholas J. Belkin 7. Beyond English - Donna K. Harman 8. Retrieving Noisy Text - Ellen M. Voorhees and John S. Garofolo 9.The Very Large Collection and Web Tracks - David Hawking and Nick Craswell 10. Question Answering in TREC - Ellen M. Voorhees 11. The University of Massachusetts and a Dozen TRECs - James Allan, W. Bruce Croft and Jamie Callan 12. How Okapi Came to TREC - Stephen Robertson 13. The SMART Project at TREC - Chris Buckley 14. Ten Years of Ad Hoc Retrieval at TREC Using PIRCS - Kui-Lam Kwok 15. MultiText Experiments for TREC - Gordon V. Cormack, Charles L. A. Clarke, Christopher R. Palmer and Thomas R. Lynam 16. A Language-Modeling Approach to TREC - Djoerd Hiemstra and Wessel Kraaij 17. BM Research Activities at TREC - Eric W. Brown, David Carmel, Martin Franz, Abraham Ittycheriah, Tapas Kanungo, Yoelle Maarek, J. Scott McCarley, Robert L. Mack, John M. Prager, John R. Smith, Aya Soffer, Jason Y. Zien and Alan D. Marwick Epilogue: Metareflections on TREC - Karen Sparck Jones
    Footnote
    Rez. in: JASIST 58(2007) no.6, S.910-911 (J.L. Vicedo u. J. Gomez): "The Text REtrieval Conference (TREC) is a yearly workshop hosted by the U.S. government's National Institute of Standards and Technology (NIST) that fosters and supports research in information retrieval as well as speeding the transfer of technology between research labs and industry. Since 1992, TREC has provided the infrastructure necessary for large-scale evaluations of different text retrieval methodologies. TREC impact has been very important and its success has been mainly supported by its continuous adaptation to the emerging information retrieval needs. Not in vain, TREC has built evaluation benchmarks for more than 20 different retrieval problems such as Web retrieval, speech retrieval, or question-answering. The large and intense trajectory of annual TREC conferences has resulted in an immense bulk of documents reflecting the different eval uation and research efforts developed. This situation makes it difficult sometimes to observe clearly how research in information retrieval (IR) has evolved over the course of TREC. TREC: Experiment and Evaluation in Information Retrieval succeeds in organizing and condensing all this research into a manageable volume that describes TREC history and summarizes the main lessons learned. The book is organized into three parts. The first part is devoted to the description of TREC's origin and history, the test collections, and the evaluation methodology developed. The second part describes a selection of the major evaluation exercises (tracks), and the third part contains contributions from research groups that had a large and remarkable participation in TREC. Finally, Karen Spark Jones, one of the main promoters of research in IR, closes the book with an epilogue that analyzes the impact of TREC on this research field.
    ... TREC: Experiment and Evaluation in Information Retrieval is a reliable and comprehensive review of the TREC program and has been adopted by NIST as the official history of TREC (see http://trec.nist.gov). We were favorably surprised by the book. Well structured and written, chapters are self-contained and the existence of references to specialized and more detailed publications is continuous, which makes it easier to expand into the different aspects analyzed in the text. This book succeeds in compiling TREC evolution from its inception in 1992 to 2003 in an adequate and manageable volume. Thanks to the impressive effort performed by the authors and their experience in the field, it can satiate the interests of a great variety of readers. While expert researchers in the IR field and IR-related industrial companies can use it as a reference manual, it seems especially useful for students and non-expert readers willing to approach this research area. Like NIST, we would recommend this reading to anyone who may be interested in textual information retrieval."
    LCSH
    Information storage and retrieval systems / Congresses
    RSWK
    Information Retrieval / Textverarbeitung / Aufsatzsammlung (BVB)
    Kongress / Information Retrieval / Kongress (GBV)
    Subject
    Information Retrieval / Textverarbeitung / Aufsatzsammlung (BVB)
    Kongress / Information Retrieval / Kongress (GBV)
    Information storage and retrieval systems / Congresses
  2. Spinning the Semantic Web : bringing the World Wide Web to its full potential (2003) 0.01
    0.0054467996 = product of:
      0.013616999 = sum of:
        0.0063072164 = weight(_text_:a in 1981) [ClassicSimilarity], result of:
          0.0063072164 = score(doc=1981,freq=14.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.11796933 = fieldWeight in 1981, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1981)
        0.0073097823 = product of:
          0.014619565 = sum of:
            0.014619565 = weight(_text_:information in 1981) [ClassicSimilarity], result of:
              0.014619565 = score(doc=1981,freq=14.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.1796046 = fieldWeight in 1981, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1981)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    As the World Wide Web continues to expand, it becomes increasingly difficult for users to obtain information efficiently. Because most search engines read format languages such as HTML or SGML, search results reflect formatting tags more than actual page content, which is expressed in natural language. Spinning the Semantic Web describes an exciting new type of hierarchy and standardization that will replace the current "Web of links" with a "Web of meaning." Using a flexible set of languages and tools, the Semantic Web will make all available information - display elements, metadata, services, images, and especially content - accessible. The result will be an immense repository of information accessible for a wide range of new applications. This first handbook for the Semantic Web covers, among other topics, software agents that can negotiate and collect information, markup languages that can tag many more types of information in a document, and knowledge systems that enable machines to read Web pages and determine their reliability. The truly interdisciplinary Semantic Web combines aspects of artificial intelligence, markup languages, natural language processing, information retrieval, knowledge representation, intelligent agents, and databases.
    Content
    Inhalt: Tim Bemers-Lee: The Original Dream - Re-enter Machines - Where Are We Now? - The World Wide Web Consortium - Where Is the Web Going Next? / Dieter Fensel, James Hendler, Henry Lieberman, and Wolfgang Wahlster: Why Is There a Need for the Semantic Web and What Will It Provide? - How the Semantic Web Will Be Possible / Jeff Heflin, James Hendler, and Sean Luke: SHOE: A Blueprint for the Semantic Web / Deborah L. McGuinness, Richard Fikes, Lynn Andrea Stein, and James Hendler: DAML-ONT: An Ontology Language for the Semantic Web / Michel Klein, Jeen Broekstra, Dieter Fensel, Frank van Harmelen, and Ian Horrocks: Ontologies and Schema Languages on the Web / Borys Omelayenko, Monica Crubezy, Dieter Fensel, Richard Benjamins, Bob Wielinga, Enrico Motta, Mark Musen, and Ying Ding: UPML: The Language and Tool Support for Making the Semantic Web Alive / Deborah L. McGuinness: Ontologies Come of Age / Jeen Broekstra, Arjohn Kampman, and Frank van Harmelen: Sesame: An Architecture for Storing and Querying RDF Data and Schema Information / Rob Jasper and Mike Uschold: Enabling Task-Centered Knowledge Support through Semantic Markup / Yolanda Gil: Knowledge Mobility: Semantics for the Web as a White Knight for Knowledge-Based Systems / Sanjeev Thacker, Amit Sheth, and Shuchi Patel: Complex Relationships for the Semantic Web / Alexander Maedche, Steffen Staab, Nenad Stojanovic, Rudi Studer, and York Sure: SEmantic portAL: The SEAL Approach / Ora Lassila and Mark Adler: Semantic Gadgets: Ubiquitous Computing Meets the Semantic Web / Christopher Frye, Mike Plusch, and Henry Lieberman: Static and Dynamic Semantics of the Web / Masahiro Hori: Semantic Annotation for Web Content Adaptation / Austin Tate, Jeff Dalton, John Levine, and Alex Nixon: Task-Achieving Agents on the World Wide Web
  3. Floridi, L.: Philosophy and computing : an introduction (1999) 0.01
    0.005278751 = product of:
      0.013196876 = sum of:
        0.0076151006 = weight(_text_:a in 823) [ClassicSimilarity], result of:
          0.0076151006 = score(doc=823,freq=10.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.14243183 = fieldWeight in 823, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=823)
        0.0055817757 = product of:
          0.011163551 = sum of:
            0.011163551 = weight(_text_:information in 823) [ClassicSimilarity], result of:
              0.011163551 = score(doc=823,freq=4.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.13714671 = fieldWeight in 823, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=823)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Philosophy and Computing explores each of the following areas of technology: the digital revolution; the computer; the Internet and the Web; CD-ROMs and Mulitmedia; databases, textbases, and hypertexts; Artificial Intelligence; the future of computing. Luciano Floridi shows us how the relationship between philosophy and computing provokes a wide range of philosophical questions: is there a philosophy of information? What can be achieved by a classic computer? How can we define complexity? What are the limits of quantam computers? Is the Internet an intellectual space or a polluted environment? What is the paradox in the Strong Artificial Intlligence program? Philosophy and Computing is essential reading for anyone wishing to fully understand both the development and history of information and communication technology as well as the philosophical issues it ultimately raises. 'The most careful and scholarly book to be written on castles in a generation.'
  4. Farkas, M.G.: Social software in libraries : building collaboration, communication, and community online (2007) 0.00
    0.0049160775 = product of:
      0.012290194 = sum of:
        0.004086692 = weight(_text_:a in 2364) [ClassicSimilarity], result of:
          0.004086692 = score(doc=2364,freq=2.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.07643694 = fieldWeight in 2364, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=2364)
        0.008203502 = product of:
          0.016407004 = sum of:
            0.016407004 = weight(_text_:information in 2364) [ClassicSimilarity], result of:
              0.016407004 = score(doc=2364,freq=6.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.20156369 = fieldWeight in 2364, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2364)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Content
    Inhalt: What is social software? -- Blogs -- Blogs in libraries : practical applications -- RSS -- Wikis -- Online communities -- Social networking -- Social bookmarking and collaborative filtering -- Tools for synchronous online reference -- The mobile revolution -- Podcasting -- Screencasting and vodcasting -- Gaming -- What will work @ your library -- Keeping up : a primer -- Future trends in social software.
    Imprint
    Medford, N.J. : Information Today
    LCSH
    Libraries / Information technology
    Subject
    Libraries / Information technology
  5. Hars, A.: From publishing to knowledge networks : reinventing online knowledge infrastructures (2003) 0.00
    0.0048924554 = product of:
      0.012231139 = sum of:
        0.0034055763 = weight(_text_:a in 1634) [ClassicSimilarity], result of:
          0.0034055763 = score(doc=1634,freq=2.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.06369744 = fieldWeight in 1634, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1634)
        0.008825562 = product of:
          0.017651124 = sum of:
            0.017651124 = weight(_text_:information in 1634) [ClassicSimilarity], result of:
              0.017651124 = score(doc=1634,freq=10.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.21684799 = fieldWeight in 1634, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1634)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    LCSH
    Information technology
    RSWK
    Information Engineering / Wissensmanagement (BVB)
    Subject
    Information Engineering / Wissensmanagement (BVB)
    Information technology
    Theme
    Information Resources Management
  6. Research and advanced technology for digital libraries : 8th European conference, ECDL 2004, Bath, UK, September 12-17, 2004 : proceedings (2004) 0.00
    0.0048828027 = product of:
      0.012207007 = sum of:
        0.00385297 = weight(_text_:a in 2427) [ClassicSimilarity], result of:
          0.00385297 = score(doc=2427,freq=4.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.072065435 = fieldWeight in 2427, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.03125 = fieldNorm(doc=2427)
        0.008354037 = product of:
          0.016708074 = sum of:
            0.016708074 = weight(_text_:information in 2427) [ClassicSimilarity], result of:
              0.016708074 = score(doc=2427,freq=14.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.20526241 = fieldWeight in 2427, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2427)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    This book constitutes the refereed proceedings of the 8th European Conference on Research and Advanced Technology for Digital Libraries, ECDL 2004, held in Bath, UK in September 2004. The 47 revised full papers presented were carefully reviewed and selected from a total of 148 submissions. The papers are organized in topical sections on digital library architectures, evaluation and usability, user interfaces and presentation, new approaches to information retrieval, interoperability, enhanced indexing and search methods, personalization and applications, music digital libraries, personal digital libraries, innovative technologies, open archive initiative, new models and tools, and user-centered design.
    Content
    Inhalt: Digital Library Architectures Evaluation and Usability User Interfaces and Presentation New Approaches to Information Retrieval - From Abstract to Virtual Entities: Implementation of Work-Based Searching in a Multimedia Digital Library / Mark Notess, Jenn Riley, and Harriette Hemmasi Interoperability Enhanced Indexing and Searching Methods Personalisation and Annotation Music Digital Libraries Personal Digital Libraries Innovative Technologies for Digital Libraries Open Archives Initiative New Models and Tools User-Centred Design - Evaluating Strategic Support for Information Access in the DAFFODIL System / Claus-Peter Klas, Norbert Fuhr, and Andre Schaefer Innovative Technologies for Digital Libraries
    RSWK
    Elektronische Bibliothek / Multimedia / Information Retrieval / Kongress / Bath <2004>
    World Wide Web / Elektronische Bibliothek / Information Retrieval / Kongress / Bath <2004>
    Subject
    Elektronische Bibliothek / Multimedia / Information Retrieval / Kongress / Bath <2004>
    World Wide Web / Elektronische Bibliothek / Information Retrieval / Kongress / Bath <2004>
  7. New directions in cognitive information retrieval (2005) 0.00
    0.004825914 = product of:
      0.012064785 = sum of:
        0.004170962 = weight(_text_:a in 338) [ClassicSimilarity], result of:
          0.004170962 = score(doc=338,freq=12.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.07801312 = fieldWeight in 338, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.01953125 = fieldNorm(doc=338)
        0.007893822 = product of:
          0.015787644 = sum of:
            0.015787644 = weight(_text_:information in 338) [ClassicSimilarity], result of:
              0.015787644 = score(doc=338,freq=32.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.19395474 = fieldWeight in 338, product of:
                  5.656854 = tf(freq=32.0), with freq of:
                    32.0 = termFreq=32.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=338)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Editor
    Spink, A. u. C. Cole
    Footnote
    Rez. in: Mitt. VÖB 59(2006) H.3, S.95-98 (O. Oberhauser): "Dieser Sammelband der Herausgeber A. Spink & C. Cole ist kurz vor ihrem im letzten Heft der Mitteilungen der VÖB besprochenen zweiten Buch erschienen. Er wendet sich an Informationswissenschaftler, Bibliothekare, Sozialwissenschaftler sowie Informatiker mit Interesse am Themenbereich Mensch-Computer-Interaktion und präsentiert einen Einblick in die aktuelle Forschung zum kognitiv orientierten Information Retrieval. Diese Richtung, die von der Analyse der Informationsprobleme der Benutzer und deren kognitivem Verhalten bei der Benutzung von Informationssystemen ausgeht, steht in einem gewissen Kontrast zum traditionell vorherrschenden IR-Paradigma, das sich auf die Optimierung der IR-Systeme und ihrer Effizienz konzentriert. "Cognitive information retrieval" oder CIR (natürlich geht es auch hier nicht ohne ein weiteres Akronym ab) ist ein interdisziplinärer Forschungsbereich, der Aktivitäten aus Informationswissenschaft, Informatik, Humanwissenschaften, Kognitionswissenschaft, Mensch-Computer-Interaktion und anderen informationsbezogenen Gebieten inkludiert.
    CIR Concepts - Interactive information retrieval: Bringing the user to a selection state, von Charles Cole et al. (Montréal), konzentriert sich auf den kognitiven Aspekt von Benutzern bei der Interaktion mit den bzw. der Reaktion auf die vom IR-System ausgesandten Stimuli; "selection" bezieht sich dabei auf die Auswahl, die das System den Benutzern abverlangt und die zur Veränderung ihrer Wissensstrukturen beiträgt. - Cognitive overlaps along the polyrepresentation continuum, von Birger Larsen und Peter Ingwersen (Kopenhagen), beschreibt einen auf Ingwersens Principle of Polyrepresentation beruhenden methodischen Ansatz, der dem IR-System ein breiteres Bild des Benutzers bzw. der Dokumente vermittelt als dies bei herkömmlichen, lediglich anfragebasierten Systemen möglich ist. - Integrating approaches to relevance, von Ian Ruthven (Glasgow), analysiert den Relevanzbegriff und schlägt anstelle des gegenwärtig in IR-Systemverwendeten, eindimensionalen Relevanzkonzepts eine multidimensionale Sichtweise vor. - New cognitive directions, von Nigel Ford (Sheffield), führt neue Begriffe ein: Ford schlägt anstelle von information need und information behaviour die Alternativen knowledge need und knowledge behaviour vor.
    CIR Processes - A multitasking framework for cognitive information retrieval, von Amanda Spink und Charles Cole (Australien/Kanada), sieht - im Gegensatz zu traditionellen Ansätzen - die simultane Bearbeitung verschiedener Aufgaben (Themen) während einer Informationssuche als den Normalfall an und analysiert das damit verbundene Benutzerverhalten. - Explanation in information seeking and retrieval, von Pertti Vakkari und Kalervo Järvelin (Tampere), plädiert anhand zweier empirischer Untersuchungen für die Verwendung des aufgabenorientierten Ansatzes ("task") in der IR-Forschung, gerade auch als Bindeglied zwischen nicht ausreichend mit einander kommunizierenden Disziplinen (Informationswissenschaft, Informatik, diverse Sozialwissenschaften). - Towards an alternative information retrieval system for children, von Jamshid Beheshti et al. (Montréal), berichtet über den Stand der IR-Forschung für Kinder und schlägt vor, eine Metapher aus dem Sozialkonstruktivismus (Lernen als soziales Verhandeln) als Gestaltungsprinzip für einschlägige IR-Systeme zu verwenden. CIR Techniques - Implicit feedback: using behavior to infer relevance, von Diane Kelly (North Carolina), setzt sich kritisch mit den Techniken zur Analyse des von Benutzern von IR-Systemen geäußerten Relevance-Feedbacks - explizit und implizit - auseinander. - Educational knowledge domain visualizations, von Peter Hook und Katy Börner (Indiana), beschreibt verschiedene Visualisierungstechniken zur Repräsentation von Wissensgebieten, die "Novizen" bei der Verwendung fachspezifischer IR-Systeme unterstützen sollen. - Learning and training to search, von Wendy Lucas und Heikki Topi (Massachusetts), analysiert, im breiteren Kontext der Information- Seeking-Forschung, Techniken zur Schulung von Benutzern von IRSystemen.
    Sämtliche Beiträge sind von hohem Niveau und bieten anspruchsvolle Lektüre. Verallgemeinert formuliert, fragen sie nach der Verknüpfung zwischen dem breiteren Kontext des Warum und Wie der menschlichen Informationssuche und den technischen bzw. sonstigen Randbedingungen, die die Interaktion zwischen Benutzern und Systemen bestimmen. Natürlich liegt hier kein Hand- oder Lehrbuch vor, sodass man - fairerweise - nicht von einer systematischen Behandlung dieses Themenbereichs oder einem didaktischen Aufbau ausgehen bzw. derlei erwarten darf. Das Buch bietet jedenfalls einen guten und vielfältigen Einstieg und Einblick in dieses interessante Forschungsgebiet. Fachlich einschlägige und größere allgemeine Bibliotheken sollten es daher jedenfalls in ihren Bestand aufnehmen. Schon die Rezension des oben zitierten zweiten Buches des Herausgeber-Duos Spink-Cole enthielt einen kritischen Hinweis auf das dortige Sachregister. Der vorliegende Band erfordert noch stärkere Nerven, denn der hier als "Index" bezeichnete Seitenfüller spottet geradezu jeder Beschreibung, umso mehr, als wir uns in einem informationswissenschaftlichen Kontext befi nden. Was soll man denn tatsächlich mit Einträgen wie "information" anfangen, noch dazu, wenn dazu über 150 verschiedene Seitenzahlen angegeben werden? Ähnlich verhält es sich mit anderen allgemeinen Begriffen wie z.B. "knowledge", "model", "tasks", "use", "users" - allesamt mit einer gewaltigen Menge von Seitenzahlen versehen und damit ohne Wert! Dieses der Leserschaft wenig dienliche Register ist wohl dem Verlag anzulasten, auch wenn die Herausgeber selbst seine Urheber gewesen sein sollten. Davon abgesehen wurde wieder einmal ein solide gefertigter Band vorgelegt, der allerdings wegen seines hohen Preis eher nur institutionelle Käufer ansprechen wird."
    Weitere Rez. in: JASIST 58(2007) no.5, S.758-760 (A. Gruzd): "Despite the minor drawbacks described, the book is a great source for researchers in the IR&S fields in general and in the CIR field in particular. Furthermore, different chapters of this book also might be of interest to members from other communities. For instance, librarians responsible for library instruction might find the chapter on search training by Lucas and Topi helpful in their work. Cognitive psychologists would probably be intrigued by Spink and Cole's view on multitasking. IR interface designers will likely find the chapter on KDV by Hook and Borner very beneficial. And students taking IR-related courses might find the thorough literature reviews by Ruthven and Kelly particularly useful when beginning their own research."
    LCSH
    Information retrieval ; Human / computer interaction
    RSWK
    Kognition / Informationsverarbeitung / Information Retrieval / Aufsatzsammlung
    Series
    The information retrieval series, vol. 19
    Subject
    Kognition / Informationsverarbeitung / Information Retrieval / Aufsatzsammlung
    Information retrieval ; Human / computer interaction
  8. Broughton, V.: Essential thesaurus construction (2006) 0.00
    0.004713759 = product of:
      0.011784397 = sum of:
        0.006092081 = weight(_text_:a in 2924) [ClassicSimilarity], result of:
          0.006092081 = score(doc=2924,freq=40.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.11394546 = fieldWeight in 2924, product of:
              6.3245554 = tf(freq=40.0), with freq of:
                40.0 = termFreq=40.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.015625 = fieldNorm(doc=2924)
        0.005692316 = product of:
          0.011384632 = sum of:
            0.011384632 = weight(_text_:information in 2924) [ClassicSimilarity], result of:
              0.011384632 = score(doc=2924,freq=26.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.13986275 = fieldWeight in 2924, product of:
                  5.0990195 = tf(freq=26.0), with freq of:
                    26.0 = termFreq=26.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.015625 = fieldNorm(doc=2924)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Many information professionals working in small units today fail to find the published tools for subject-based organization that are appropriate to their local needs, whether they are archivists, special librarians, information officers, or knowledge or content managers. Large established standards for document description and organization are too unwieldy, unnecessarily detailed, or too expensive to install and maintain. In other cases the available systems are insufficient for a specialist environment, or don't bring things together in a helpful way. A purpose built, in-house system would seem to be the answer, but too often the skills necessary to create one are lacking. This practical text examines the criteria relevant to the selection of a subject-management system, describes the characteristics of some common types of subject tool, and takes the novice step by step through the process of creating a system for a specialist environment. The methodology employed is a standard technique for the building of a thesaurus that incidentally creates a compatible classification or taxonomy, both of which may be used in a variety of ways for document or information management. Key areas covered are: What is a thesaurus? Tools for subject access and retrieval; what a thesaurus is used for? Why use a thesaurus? Examples of thesauri; the structure of a thesaurus; thesaural relationships; practical thesaurus construction; the vocabulary of the thesaurus; building the systematic structure; conversion to alphabetic format; forms of entry in the thesaurus; maintaining the thesaurus; thesaurus software; and; the wider environment. Essential for the practising information professional, this guide is also valuable for students of library and information science.
    Footnote
    Rez. in: Mitt. VÖB 60(2007) H.1, S.98-101 (O. Oberhauser): "Die Autorin von Essential thesaurus construction (and essential taxonomy construction, so der implizite Untertitel, vgl. S. 1) ist durch ihre Lehrtätigkeit an der bekannten School of Library, Archive and Information Studies des University College London und durch ihre bisherigen Publikationen auf den Gebieten (Facetten-)Klassifikation und Thesaurus fachlich einschlägig ausgewiesen. Nach Essential classification liegt nun ihr Thesaurus-Lehrbuch vor, mit rund 200 Seiten Text und knapp 100 Seiten Anhang ein handliches Werk, das seine Genese zum Grossteil dem Lehrbetrieb verdankt, wie auch dem kurzen Einleitungskapitel zu entnehmen ist. Das Buch ist der Schule von Jean Aitchison et al. verpflichtet und wendet sich an "the indexer" im weitesten Sinn, d.h. an alle Personen, die ein strukturiertes, kontrolliertes Fachvokabular für die Zwecke der sachlichen Erschliessung und Suche erstellen wollen bzw. müssen. Es möchte dieser Zielgruppe das nötige methodische Rüstzeug für eine solche Aufgabe vermitteln, was einschliesslich der Einleitung und der Schlussbemerkungen in zwanzig Kapiteln geschieht - eine ansprechende Strukturierung, die ein wohldosiertes Durcharbeiten möglich macht. Zu letzterem tragen auch die von der Autorin immer wieder gestellten Übungsaufgaben bei (Lösungen jeweils am Kapitelende). Zu Beginn der Darstellung wird der "information retrieval thesaurus" von dem (zumindest im angelsächsischen Raum) weit öfter mit dem Thesaurusbegriff assoziierten "reference thesaurus" abgegrenzt, einem nach begrifflicher Ähnlichkeit angeordneten Synonymenwörterbuch, das gerne als Mittel zur stilistischen Verbesserung beim Abfassen von (wissenschaftlichen) Arbeiten verwendet wird. Ohne noch ins Detail zu gehen, werden optische Erscheinungsform und Anwendungsgebiete von Thesauren vorgestellt, der Thesaurus als postkoordinierte Indexierungssprache erläutert und seine Nähe zu facettierten Klassifikationssystemen erwähnt. In der Folge stellt Broughton die systematisch organisierten Systeme (Klassifikation/ Taxonomie, Begriffs-/Themendiagramme, Ontologien) den alphabetisch angeordneten, wortbasierten (Schlagwortlisten, thesaurusartige Schlagwortsysteme und Thesauren im eigentlichen Sinn) gegenüber, was dem Leser weitere Einordnungshilfen schafft. Die Anwendungsmöglichkeiten von Thesauren als Mittel der Erschliessung (auch als Quelle für Metadatenangaben bei elektronischen bzw. Web-Dokumenten) und der Recherche (Suchformulierung, Anfrageerweiterung, Browsing und Navigieren) kommen ebenso zur Sprache wie die bei der Verwendung natürlichsprachiger Indexierungssysteme auftretenden Probleme. Mit Beispielen wird ausdrücklich auf die mehr oder weniger starke fachliche Spezialisierung der meisten dieser Vokabularien hingewiesen, wobei auch Informationsquellen über Thesauren (z.B. www.taxonomywarehouse.com) sowie Thesauren für nicht-textuelle Ressourcen kurz angerissen werden.
    In den stärker ins Detail gehenden Kapiteln weist Broughton zunächst auf die Bedeutung des systematischen Teils eines Thesaurus neben dem alphabetischen Teil hin und erläutert dann die Elemente des letzteren, wobei neben den gängigen Thesaurusrelationen auch die Option der Ausstattung der Einträge mit Notationen eines Klassifikationssystems erwähnt wird. Die Thesaurusrelationen selbst werden später noch in einem weiteren Kapitel ausführlicher diskutiert, wobei etwa auch die polyhierarchische Beziehung thematisiert wird. Zwei Kapitel zur Vokabularkontrolle führen in Aspekte wie Behandlung von Synonymen, Vermeidung von Mehrdeutigkeit, Wahl der bevorzugten Terme sowie die Formen von Thesauruseinträgen ein (grammatische Form, Schreibweise, Zeichenvorrat, Singular/Plural, Komposita bzw. deren Zerlegung usw.). Insgesamt acht Kapitel - in der Abfolge mit den bisher erwähnten Abschnitten didaktisch geschickt vermischt - stehen unter dem Motto "Building a thesaurus". Kurz zusammengefasst, geht es dabei um folgende Tätigkeiten und Prozesse: - Sammlung des Vokabulars unter Nutzung entsprechender Quellen; - Termextraktion aus den Titeln von Dokumenten und Probleme hiebei; - Analyse des Vokabulars (Facettenmethode); - Einbau einer internen Struktur (Facetten und Sub-Facetten, Anordnung der Terme); - Erstellung einer hierarchischen Struktur und deren Repräsentation; - Zusammengesetzte Themen bzw. Begriffe (Facettenanordnung: filing order vs. citation order); - Konvertierung der taxonomischen Anordnung in ein alphabetisches Format (Auswahl der Vorzugsbegriffe, Identifizieren hierarchischer Beziehungen, verwandter Begriffe usw.); - Erzeugen der endgültigen Thesaurus-Einträge.
    In einem abschliessenden Kapitel geht das Buch auf Thesauruspflege und -verwaltung ein, wobei auch das Thema "Thesaurussoftware" angerissen wird - letzteres vielleicht ein wenig zu kurz. Erst hier mag manchem unbefangenen Leser bewusst werden, dass die in den vorhergehenden Kapiteln dargestellte Methodik eigentlich ohne den Einsatz dezidierter Software besprochen wurde, ja vielleicht auch so besprochen werden musste, um ein entsprechendes Verständnis herzustellen. In der nachfolgenden zweiseitigen Conclusio wird erwähnt, dass die britische Norm Structured Vocabularies for Information Retrieval (BS 8723) vor einer Neufassung stehe - was den Rezensenten darauf hinzuweisen gemahnt, dass sich dieses Buch natürlich ausschliesslich auf den anglo-amerikanischen Sprachraum und die dort geltenden Thesaurus-Gepflogenheiten bezieht. Der relativ umfangreiche Anhang beinhaltet ausser Materialie zum erwähnten Demonstrationsbeispiel auch ein nützliches Glossarium sowie ein professionell gefertigtes Sachregister. Literaturhinweise werden - in sparsamer Dosierung - jeweils am Ende der einzelnen Kapitel gegeben, sodass sich die "Bibliography" am Ende des Buches auf einige Normen und zwei Standardwerke beschränken kann. Realistisch betrachtet, darf vermutlich nicht davon ausgegangen werden, dass Leser, die dieses Buch durchgearbeitet haben, sogleich in der Lage sein werden, eigenständig einen Thesaurus zu erstellen. Ein Lehrbuch allein kann weder einen Kurs noch die praktische Erfahrung ersetzen, die für eine solche Tätigkeit vonnöten sind. Ich kann mir aber gut vorstellen, dass die Kenntnis der in diesem Buch vermittelten Inhalte sehr wohl in die Lage versetzt, in einem Team, das einen Thesaurus erstellen soll, kompetent mitzuarbeiten, mit den Konzepten und Fachtermini zurechtzukommen und selbst konstruktive Beiträge zu leisten. Ausserdem erscheint mir das Werk hervorragend als Begleitmaterial zu einer Lehrveranstaltung geeignet - oder auch als Grundlage für die Planung einer solchen. Ein britischer Einführungstext eben, im besten Sinne."
    Weitere Rez. in: New Library World 108(2007) nos.3/4, S.190-191 (K.V. Trickey): "Vanda has provided a very useful work that will enable any reader who is prepared to follow her instruction to produce a thesaurus that will be a quality language-based subject access tool that will make the task of information retrieval easier and more effective. Once again I express my gratitude to Vanda for producing another excellent book." - Electronic Library 24(2006) no.6, S.866-867 (A.G. Smith): "Essential thesaurus construction is an ideal instructional text, with clear bullet point summaries at the ends of sections, and relevant and up to date references, putting thesauri in context with the general theory of information retrieval. But it will also be a valuable reference for any information professional developing or using a controlled vocabulary." - KO 33(2006) no.4, S.215-216 (M.P. Satija)
    LCSH
    Information organization
    Subject
    Information organization
  9. Chu, H.: Information representation and retrieval in the digital age (2010) 0.00
    0.0046886913 = product of:
      0.023443457 = sum of:
        0.023443457 = product of:
          0.046886913 = sum of:
            0.046886913 = weight(_text_:information in 377) [ClassicSimilarity], result of:
              0.046886913 = score(doc=377,freq=36.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.5760162 = fieldWeight in 377, product of:
                  6.0 = tf(freq=36.0), with freq of:
                    36.0 = termFreq=36.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=377)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Content
    Information representation and retrieval : an overview -- Information representation I : basic approaches -- Information representation II : related topics -- Language in information representation and retrieval -- Retrieval techniques and query representation -- Retrieval approaches -- Information retrieval models -- Information retrieval systems -- Retrieval of information unique in content or format -- The user dimension in information representation and retrieval -- Evaluation of information representation and retrieval -- Artificial intelligence in information representation and retrieval.
    Imprint
    Medford, NJ : Information Today
    LCSH
    Information organization
    Information retrieval
    Information storage and retrieval systems
    Subject
    Information organization
    Information retrieval
    Information storage and retrieval systems
  10. Survey of text mining : clustering, classification, and retrieval (2004) 0.00
    0.0045197597 = product of:
      0.011299399 = sum of:
        0.0034055763 = weight(_text_:a in 804) [ClassicSimilarity], result of:
          0.0034055763 = score(doc=804,freq=2.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.06369744 = fieldWeight in 804, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=804)
        0.007893822 = product of:
          0.015787644 = sum of:
            0.015787644 = weight(_text_:information in 804) [ClassicSimilarity], result of:
              0.015787644 = score(doc=804,freq=8.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.19395474 = fieldWeight in 804, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=804)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Extracting content from text continues to be an important research problem for information processing and management. Approaches to capture the semantics of text-based document collections may be based on Bayesian models, probability theory, vector space models, statistical models, or even graph theory. As the volume of digitized textual media continues to grow, so does the need for designing robust, scalable indexing and search strategies (software) to meet a variety of user needs. Knowledge extraction or creation from text requires systematic yet reliable processing that can be codified and adapted for changing needs and environments. This book will draw upon experts in both academia and industry to recommend practical approaches to the purification, indexing, and mining of textual information. It will address document identification, clustering and categorizing documents, cleaning text, and visualizing semantic models of text.
    LCSH
    Data mining ; Information retrieval
    Subject
    Data mining ; Information retrieval
  11. Berry, M.W.; Browne, M.: Understanding search engines : mathematical modeling and text retrieval (2005) 0.00
    0.0044313995 = product of:
      0.011078498 = sum of:
        0.002724461 = weight(_text_:a in 7) [ClassicSimilarity], result of:
          0.002724461 = score(doc=7,freq=2.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.050957955 = fieldWeight in 7, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.03125 = fieldNorm(doc=7)
        0.008354037 = product of:
          0.016708074 = sum of:
            0.016708074 = weight(_text_:information in 7) [ClassicSimilarity], result of:
              0.016708074 = score(doc=7,freq=14.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.20526241 = fieldWeight in 7, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.03125 = fieldNorm(doc=7)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    The second edition of Understanding Search Engines: Mathematical Modeling and Text Retrieval follows the basic premise of the first edition by discussing many of the key design issues for building search engines and emphasizing the important role that applied mathematics can play in improving information retrieval. The authors discuss important data structures, algorithms, and software as well as user-centered issues such as interfaces, manual indexing, and document preparation. Significant changes bring the text up to date on current information retrieval methods: for example the addition of a new chapter on link-structure algorithms used in search engines such as Google. The chapter on user interface has been rewritten to specifically focus on search engine usability. In addition the authors have added new recommendations for further reading and expanded the bibliography, and have updated and streamlined the index to make it more reader friendly.
    Content
    Inhalt: Introduction Document File Preparation - Manual Indexing - Information Extraction - Vector Space Modeling - Matrix Decompositions - Query Representations - Ranking and Relevance Feedback - Searching by Link Structure - User Interface - Book Format Document File Preparation Document Purification and Analysis - Text Formatting - Validation - Manual Indexing - Automatic Indexing - Item Normalization - Inverted File Structures - Document File - Dictionary List - Inversion List - Other File Structures Vector Space Models Construction - Term-by-Document Matrices - Simple Query Matching - Design Issues - Term Weighting - Sparse Matrix Storage - Low-Rank Approximations Matrix Decompositions QR Factorization - Singular Value Decomposition - Low-Rank Approximations - Query Matching - Software - Semidiscrete Decomposition - Updating Techniques Query Management Query Binding - Types of Queries - Boolean Queries - Natural Language Queries - Thesaurus Queries - Fuzzy Queries - Term Searches - Probabilistic Queries Ranking and Relevance Feedback Performance Evaluation - Precision - Recall - Average Precision - Genetic Algorithms - Relevance Feedback Searching by Link Structure HITS Method - HITS Implementation - HITS Summary - PageRank Method - PageRank Adjustments - PageRank Implementation - PageRank Summary User Interface Considerations General Guidelines - Search Engine Interfaces - Form Fill-in - Display Considerations - Progress Indication - No Penalties for Error - Results - Test and Retest - Final Considerations Further Reading
    RSWK
    Suchmaschine / Information Retrieval
    Suchmaschine / Information Retrieval / Mathematisches Modell (HEBIS)
    Subject
    Suchmaschine / Information Retrieval
    Suchmaschine / Information Retrieval / Mathematisches Modell (HEBIS)
  12. Research and advanced technology for digital libraries : 9th European conference, ECDL 2005, Vienna, Austria, September 18 - 23, 2005 ; proceedings (2005) 0.00
    0.0044135856 = product of:
      0.011033963 = sum of:
        0.0047189053 = weight(_text_:a in 2423) [ClassicSimilarity], result of:
          0.0047189053 = score(doc=2423,freq=6.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.088261776 = fieldWeight in 2423, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.03125 = fieldNorm(doc=2423)
        0.006315058 = product of:
          0.012630116 = sum of:
            0.012630116 = weight(_text_:information in 2423) [ClassicSimilarity], result of:
              0.012630116 = score(doc=2423,freq=8.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.1551638 = fieldWeight in 2423, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2423)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    This book constitutes the refereed proceedings of the 9th European Conference on Research and Advanced Technology for Digital Libraries, ECDL 2005, held in Vienna, Austria in September 2005. The 41 revised full papers presented together with 2 panel papers and 30 revised poster papers were carefully reviewed and selected from a total of 162 submissions. The papers are organized in topical sections on digital library models and architectures, multimedia and hypermedia digital libraries, XML, building digital libraries, user studies, digital preservation, metadata, digital libraries and e-learning, text classification in digital libraries, searching, and text digital libraries.
    Content
    Inhalt u.a.: - Digital Library Models and Architectures - Multimedia and Hypermedia Digital Libraries - XML - Building Digital Libraries - User Studies - Digital Preservation - Metadata - Digital Libraries and e-Learning - Text Classification in Digital Libraries - Searching - - Focused Crawling Using Latent Semantic Indexing - An Application for Vertical Search Engines / George Almpanidis, Constantine Kotropoulos, Ioannis Pitas - - Active Support for Query Formulation in Virtual Digital Libraries: A Case Study with DAFFODIL / Andre Schaefer, Matthias Jordan, Claus-Peter Klas, Norbert Fuhr - - Expression of Z39.50 Supported Search Capabilities by Applying Formal Descriptions / Michalis Sfakakis, Sarantos Kapidakis - Text Digital Libraries
    Editor
    Rauber, A. et.al.
    RSWK
    Elektronische Bibliothek / Multimedia / Information Retrieval / Kongress / Wien <2005>
    World Wide Web / Elektronische Bibliothek / Information Retrieval / Kongress / Wien <2005>
    Subject
    Elektronische Bibliothek / Multimedia / Information Retrieval / Kongress / Wien <2005>
    World Wide Web / Elektronische Bibliothek / Information Retrieval / Kongress / Wien <2005>
  13. Semantic digital libraries (2009) 0.00
    0.004075163 = product of:
      0.010187907 = sum of:
        0.0047189053 = weight(_text_:a in 3371) [ClassicSimilarity], result of:
          0.0047189053 = score(doc=3371,freq=6.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.088261776 = fieldWeight in 3371, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.03125 = fieldNorm(doc=3371)
        0.0054690014 = product of:
          0.010938003 = sum of:
            0.010938003 = weight(_text_:information in 3371) [ClassicSimilarity], result of:
              0.010938003 = score(doc=3371,freq=6.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.1343758 = fieldWeight in 3371, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3371)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Libraries have always been an inspiration for the standards and technologies developed by semantic web activities. However, except for the Dublin Core specification, semantic web and social networking technologies have not been widely adopted and further developed by major digital library initiatives and projects. Yet semantic technologies offer a new level of flexibility, interoperability, and relationships for digital repositories. Kruk and McDaniel present semantic web-related aspects of current digital library activities, and introduce their functionality; they show examples ranging from general architectural descriptions to detailed usages of specific ontologies, and thus stimulate the awareness of researchers, engineers, and potential users of those technologies. Their presentation is completed by chapters on existing prototype systems such as JeromeDL, BRICKS, and Greenstone, as well as a look into the possible future of semantic digital libraries. This book is aimed at researchers and graduate students in areas like digital libraries, the semantic web, social networks, and information retrieval. This audience will benefit from detailed descriptions of both today's possibilities and also the shortcomings of applying semantic web technologies to large digital repositories of often unstructured data.
    Content
    Inhalt: Introduction to Digital Libraries and Semantic Web: Introduction / Bill McDaniel and Sebastian Ryszard Kruk - Digital Libraries and Knowledge Organization / Dagobert Soergel - Semantic Web and Ontologies / Marcin Synak, Maciej Dabrowski and Sebastian Ryszard Kruk - Social Semantic Information Spaces / John G. Breslin A Vision of Semantic Digital Libraries: Goals of Semantic Digital Libraries / Sebastian Ryszard Kruk and Bill McDaniel - Architecture of Semantic Digital Libraries / Sebastian Ryszard Kruk, Adam Westerki and Ewelina Kruk - Long-time Preservation / Markus Reis Ontologies for Semantic Digital Libraries: Bibliographic Ontology / Maciej Dabrowski, Macin Synak and Sebastian Ryszard Kruk - Community-aware Ontologies / Slawomir Grzonkowski, Sebastian Ryszard Kruk, Adam Gzella, Jakub Demczuk and Bill McDaniel Prototypes of Semantic Digital Libraries: JeromeDL: The Social Semantic Digital Library / Sebastian Ryszard Kruk, Mariusz Cygan, Adam Gzella, Tomasz Woroniecki and Maciej Dabrowski - The BRICKS Digital Library Infrastructure / Bernhard Haslhofer and Predrag Knezevié - Semantics in Greenstone / Annika Hinze, George Buchanan, David Bainbridge and Ian Witten Building the Future - Semantic Digital Libraries in Use: Hyperbooks / Gilles Falquet, Luka Nerima and Jean-Claude Ziswiler - Semantic Digital Libraries for Archiving / Bill McDaniel - Evaluation of Semantic and Social Technologies for Digital Libraries / Sebastian Ryszard Kruk, Ewelina Kruk and Katarzyna Stankiewicz - Conclusions: The Future of Semantic Digital Libraries / Sebastian Ryszard Kruk and Bill McDaniel
    Theme
    Information Gateway
  14. Borlund, P.: Evaluation of interactive information retrieval systems (2000) 0.00
    0.003789035 = product of:
      0.018945174 = sum of:
        0.018945174 = product of:
          0.03789035 = sum of:
            0.03789035 = weight(_text_:information in 2556) [ClassicSimilarity], result of:
              0.03789035 = score(doc=2556,freq=18.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.46549135 = fieldWeight in 2556, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2556)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    LCSH
    Information storage and retrieval systems / Evaluation
    RSWK
    Information Retrieval / Datenbankverwaltung / Hochschulschrift (GBV)
    Information Retrieval / Dialogsystem (SWB)
    Information Retrieval / Dialogsystem / Leistungsbewertung (BVB)
    Subject
    Information Retrieval / Datenbankverwaltung / Hochschulschrift (GBV)
    Information Retrieval / Dialogsystem (SWB)
    Information Retrieval / Dialogsystem / Leistungsbewertung (BVB)
    Information storage and retrieval systems / Evaluation
  15. Langville, A.N.; Meyer, C.D.: Google's PageRank and beyond : the science of search engine rankings (2006) 0.00
    0.003752812 = product of:
      0.00938203 = sum of:
        0.004086692 = weight(_text_:a in 6) [ClassicSimilarity], result of:
          0.004086692 = score(doc=6,freq=8.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.07643694 = fieldWeight in 6, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0234375 = fieldNorm(doc=6)
        0.0052953376 = product of:
          0.010590675 = sum of:
            0.010590675 = weight(_text_:information in 6) [ClassicSimilarity], result of:
              0.010590675 = score(doc=6,freq=10.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.1301088 = fieldWeight in 6, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=6)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Content
    Inhalt: Chapter 1. Introduction to Web Search Engines: 1.1 A Short History of Information Retrieval - 1.2 An Overview of Traditional Information Retrieval - 1.3 Web Information Retrieval Chapter 2. Crawling, Indexing, and Query Processing: 2.1 Crawling - 2.2 The Content Index - 2.3 Query Processing Chapter 3. Ranking Webpages by Popularity: 3.1 The Scene in 1998 - 3.2 Two Theses - 3.3 Query-Independence Chapter 4. The Mathematics of Google's PageRank: 4.1 The Original Summation Formula for PageRank - 4.2 Matrix Representation of the Summation Equations - 4.3 Problems with the Iterative Process - 4.4 A Little Markov Chain Theory - 4.5 Early Adjustments to the Basic Model - 4.6 Computation of the PageRank Vector - 4.7 Theorem and Proof for Spectrum of the Google Matrix Chapter 5. Parameters in the PageRank Model: 5.1 The a Factor - 5.2 The Hyperlink Matrix H - 5.3 The Teleportation Matrix E Chapter 6. The Sensitivity of PageRank; 6.1 Sensitivity with respect to alpha - 6.2 Sensitivity with respect to H - 6.3 Sensitivity with respect to vT - 6.4 Other Analyses of Sensitivity - 6.5 Sensitivity Theorems and Proofs Chapter 7. The PageRank Problem as a Linear System: 7.1 Properties of (I - alphaS) - 7.2 Properties of (I - alphaH) - 7.3 Proof of the PageRank Sparse Linear System Chapter 8. Issues in Large-Scale Implementation of PageRank: 8.1 Storage Issues - 8.2 Convergence Criterion - 8.3 Accuracy - 8.4 Dangling Nodes - 8.5 Back Button Modeling
    Chapter 9. Accelerating the Computation of PageRank: 9.1 An Adaptive Power Method - 9.2 Extrapolation - 9.3 Aggregation - 9.4 Other Numerical Methods Chapter 10. Updating the PageRank Vector: 10.1 The Two Updating Problems and their History - 10.2 Restarting the Power Method - 10.3 Approximate Updating Using Approximate Aggregation - 10.4 Exact Aggregation - 10.5 Exact vs. Approximate Aggregation - 10.6 Updating with Iterative Aggregation - 10.7 Determining the Partition - 10.8 Conclusions Chapter 11. The HITS Method for Ranking Webpages: 11.1 The HITS Algorithm - 11.2 HITS Implementation - 11.3 HITS Convergence - 11.4 HITS Example - 11.5 Strengths and Weaknesses of HITS - 11.6 HITS's Relationship to Bibliometrics - 11.7 Query-Independent HITS - 11.8 Accelerating HITS - 11.9 HITS Sensitivity Chapter 12. Other Link Methods for Ranking Webpages: 12.1 SALSA - 12.2 Hybrid Ranking Methods - 12.3 Rankings based on Traffic Flow Chapter 13. The Future of Web Information Retrieval: 13.1 Spam - 13.2 Personalization - 13.3 Clustering - 13.4 Intelligent Agents - 13.5 Trends and Time-Sensitive Search - 13.6 Privacy and Censorship - 13.7 Library Classification Schemes - 13.8 Data Fusion Chapter 14. Resources for Web Information Retrieval: 14.1 Resources for Getting Started - 14.2 Resources for Serious Study Chapter 15. The Mathematics Guide: 15.1 Linear Algebra - 15.2 Perron-Frobenius Theory - 15.3 Markov Chains - 15.4 Perron Complementation - 15.5 Stochastic Complementation - 15.6 Censoring - 15.7 Aggregation - 15.8 Disaggregation
  16. Information: Droge, Ware oder Commons? : Wertschöpfungs- und Transformationsprozesse auf den Informationsmärkten ; Proceedings des 11. Internationalen Symposiums für Informationswissenschaft (ISI 2009) ; Konstanz, 1. - 3. April 2009 / [Hochschulverband für Informationswissenschaft (HI) e.V., Konstanz] (2009) 0.00
    0.003424719 = product of:
      0.008561797 = sum of:
        0.0023839036 = weight(_text_:a in 2930) [ClassicSimilarity], result of:
          0.0023839036 = score(doc=2930,freq=2.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.044588212 = fieldWeight in 2930, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.02734375 = fieldNorm(doc=2930)
        0.006177894 = product of:
          0.012355788 = sum of:
            0.012355788 = weight(_text_:information in 2930) [ClassicSimilarity], result of:
              0.012355788 = score(doc=2930,freq=10.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.1517936 = fieldWeight in 2930, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=2930)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    BK
    06.00 / Information und Dokumentation: Allgemeines
    Classification
    06.00 / Information und Dokumentation: Allgemeines
    Content
    - Recherche und Web Rita Strebe: Empirische Untersuchung von emotionalen Reaktionen im Prozess der Informationsrecherche im Web Teresa Märt!, Christa Womser-Hacker, Thomas Mandl: Mehrsprachige Suche in Social-Tagging-Systemen Christian Maaß, Dirk Lewandowski: Frage-Antwort-Dienste als alternativer Suchansatz? Jürgen Reischer: EXCERPT - a Within-Document Retrieval System Using Summarization Techniques - Fachportale - Open Access I Stefan Baerisch, Peter Mutschke, Maximilian Stempfliuber: Informationstechnologische Aspekte der Heterogenitätsbehandlung in Fachportalen Doris Bambey: Open Access in der Erziehungswissenschaft - Voraussetzungen und Modelle der Funktionsteilung und der Verwertung von Wissen Patrick Lay: Integration heterogener Anwendungen in Fachportalen am Beispiel Sowiport Martin Uhl, Erich Weichselgartner: Aufbau einer innovativen Publikations-Infrastrukrur für die europäische Psychologie
    RSWK
    World Wide Web / Information Retrieval / Kongress / Konstanz <2009>
    Subject
    World Wide Web / Information Retrieval / Kongress / Konstanz <2009>
  17. Weller, K.: Knowledge representation in the Social Semantic Web (2010) 0.00
    0.00326269 = product of:
      0.008156725 = sum of:
        0.0033713488 = weight(_text_:a in 4515) [ClassicSimilarity], result of:
          0.0033713488 = score(doc=4515,freq=4.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.06305726 = fieldWeight in 4515, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4515)
        0.004785376 = product of:
          0.009570752 = sum of:
            0.009570752 = weight(_text_:information in 4515) [ClassicSimilarity], result of:
              0.009570752 = score(doc=4515,freq=6.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.11757882 = fieldWeight in 4515, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=4515)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    The main purpose of this book is to sum up the vital and highly topical research issue of knowledge representation on the Web and to discuss novel solutions by combining benefits of folksonomies and Web 2.0 approaches with ontologies and semantic technologies. This book contains an overview of knowledge representation approaches in past, present and future, introduction to ontologies, Web indexing and in first case the novel approaches of developing ontologies. This title combines aspects of knowledge representation for both the Semantic Web (ontologies) and the Web 2.0 (folksonomies). Currently there is no monographic book which provides a combined overview over these topics. focus on the topic of using knowledge representation methods for document indexing purposes. For this purpose, considerations from classical librarian interests in knowledge representation (thesauri, classification schemes etc.) are included, which are not part of most other books which have a stronger background in computer science.
    LCSH
    Knowledge representation (Information theory)
    Series
    Knowledge and information; vol.3
    Subject
    Knowledge representation (Information theory)
  18. Berry, M.W.; Browne, M.: Understanding search engines : mathematical modeling and text retrieval (1999) 0.00
    0.0023203006 = product of:
      0.011601503 = sum of:
        0.011601503 = product of:
          0.023203006 = sum of:
            0.023203006 = weight(_text_:information in 5777) [ClassicSimilarity], result of:
              0.023203006 = score(doc=5777,freq=12.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.2850541 = fieldWeight in 5777, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5777)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Abstract
    This book discusses many of the key design issues for building search engines and emphazises the important role that applied mathematics can play in improving information retrieval. The authors discuss not only important data structures, algorithms, and software but also user-centered issues such as interfaces, manual indexing, and document preparation. They also present some of the current problems in information retrieval that many not be familiar to applied mathematicians and computer scientists and some of the driving computational methods (SVD, SDD) for automated conceptual indexing
    RSWK
    Suchmaschine / Information Retrieval
    Suchmaschine / Information Retrieval / Mathematisches Modell (HEBIS)
    Subject
    Suchmaschine / Information Retrieval
    Suchmaschine / Information Retrieval / Mathematisches Modell (HEBIS)
  19. Stuckenschmidt, H.; Harmelen, F. van: Information sharing on the semantic web (2005) 0.00
    0.0022327104 = product of:
      0.011163551 = sum of:
        0.011163551 = product of:
          0.022327103 = sum of:
            0.022327103 = weight(_text_:information in 2789) [ClassicSimilarity], result of:
              0.022327103 = score(doc=2789,freq=16.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.27429342 = fieldWeight in 2789, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2789)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Classification
    ST 515 Informatik / Monographien / Einzelne Anwendungen der Datenverarbeitung / Wirtschaftsinformatik / Wissensmanagement, Information engineering
    LCSH
    Ontologies (Information retrieval)
    RSWK
    Semantic Web / Ontologie <Wissensverarbeitung> / Information Retrieval / Verteilung / Metadaten / Datenintegration
    RVK
    ST 515 Informatik / Monographien / Einzelne Anwendungen der Datenverarbeitung / Wirtschaftsinformatik / Wissensmanagement, Information engineering
    Series
    Advanced information and knowledge processing
    Subject
    Semantic Web / Ontologie <Wissensverarbeitung> / Information Retrieval / Verteilung / Metadaten / Datenintegration
    Ontologies (Information retrieval)
  20. Garlock, K.L.; Piontek, S.: Designing Web interfaces to library services and resources (1999) 0.00
    0.0022102704 = product of:
      0.011051352 = sum of:
        0.011051352 = product of:
          0.022102704 = sum of:
            0.022102704 = weight(_text_:information in 1550) [ClassicSimilarity], result of:
              0.022102704 = score(doc=1550,freq=8.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.27153665 = fieldWeight in 1550, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1550)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    LCSH
    Library information networks
    Library information networks / United States
    Subject
    Library information networks
    Library information networks / United States

Subjects

Classifications