Search (5 results, page 1 of 1)

  • × classification_ss:"ST 205"
  1. Manning, C.D.; Raghavan, P.; Schütze, H.: Introduction to information retrieval (2008) 0.02
    0.01596949 = product of:
      0.06387796 = sum of:
        0.06387796 = weight(_text_:term in 4041) [ClassicSimilarity], result of:
          0.06387796 = score(doc=4041,freq=4.0), product of:
            0.21904005 = queryWeight, product of:
              4.66603 = idf(docFreq=1130, maxDocs=44218)
              0.04694356 = queryNorm
            0.29162687 = fieldWeight in 4041, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.66603 = idf(docFreq=1130, maxDocs=44218)
              0.03125 = fieldNorm(doc=4041)
      0.25 = coord(1/4)
    
    Content
    Inhalt: Boolean retrieval - The term vocabulary & postings lists - Dictionaries and tolerant retrieval - Index construction - Index compression - Scoring, term weighting & the vector space model - Computing scores in a complete search system - Evaluation in information retrieval - Relevance feedback & query expansion - XML retrieval - Probabilistic information retrieval - Language models for information retrieval - Text classification & Naive Bayes - Vector space classification - Support vector machines & machine learning on documents - Flat clustering - Hierarchical clustering - Matrix decompositions & latent semantic indexing - Web search basics - Web crawling and indexes - Link analysis Vgl. die digitale Fassung unter: http://nlp.stanford.edu/IR-book/pdf/irbookprint.pdf.
  2. Rogers, R.: Information politics on the Web (2004) 0.01
    0.008992833 = product of:
      0.035971332 = sum of:
        0.035971332 = weight(_text_:frequency in 442) [ClassicSimilarity], result of:
          0.035971332 = score(doc=442,freq=2.0), product of:
            0.27643865 = queryWeight, product of:
              5.888745 = idf(docFreq=332, maxDocs=44218)
              0.04694356 = queryNorm
            0.1301241 = fieldWeight in 442, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.888745 = idf(docFreq=332, maxDocs=44218)
              0.015625 = fieldNorm(doc=442)
      0.25 = coord(1/4)
    
    Footnote
    In Chapter 5, the "Election Issue Tracker" is introduced. The Election Issue Tracker calculates currency that is defined as "frequency of mentions of the issue terms per newspaper and across newspapers" in the three major national newspapers. The Election Issue Tracker is used to study which issues resonate with the press and which do not. As one would expect, Rogers found that not all issues that are considered important or central to a political party resonate with the press. This book contains a wealth of information that can be accessed by both researcher and practitioner. Even more interesting is the fact that researchers from a wide assortment of disciplines, from political science to information science and even communication studies, will appreciate the research and insights put forth by Rogers. Concepts presented in each chapter are thoroughly described using a wide variety of cases. Albeit all the cases are of a European flavor, mainly Dutch, they are interesting and thought-provoking. I found the descriptions of Rogers various information instruments to be very interesting. Researchers can gain from an examination of these instruments as it points to an interesting method for studying activities and behaviors on the Internet. In addition, each chapter has adequate illustrations and the bibliography is comprehensive. This book will make for an ideal supplementary text for graduate courses in information science, communication and media studies, and even political science. Like all books, however, this book had its share of shortcomings. While I was able to appreciate the content of the book, and certainly commend Rogers for studying an issue of immense significance, I found the book to be very difficult to read and parse through. The book is laden with jargon, political statements, and even has several instances of deficient writing. The book also lacked a sense of structure, and this affected the presentation of Rogers' material. I would have also hoped to see some recommendations by Rogers in terms of how should researchers further the ideas he has put forth. Areas of future research, methods for studying future problems, and even insights on what the future might hold for information politics were not given enough attention in the book; in my opinion, this was a major shortcoming. Overall, I commend Rogers for putting forth a very informative book on the issues of information politics on the Web. Information politics, especially when delivered on the communication technologies such as the Web, is going to play a vital role in our societies for a long time to come. Debates will range from the politics of how information is searched for and displayed on the Web to how the Web is used to manipulate or politicize information to meet the agendas of various entities. Richard Rogers' book will be of the seminal and foundational readings on the topic for any curious minds that want to explore these issues."
  3. Rosenfeld, L.; Morville, P.: Information architecture for the World Wide Web : designing large-scale Web sites (2007) 0.00
    0.003975128 = product of:
      0.015900511 = sum of:
        0.015900511 = product of:
          0.031801023 = sum of:
            0.031801023 = weight(_text_:22 in 5135) [ClassicSimilarity], result of:
              0.031801023 = score(doc=5135,freq=2.0), product of:
                0.16438834 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04694356 = queryNorm
                0.19345059 = fieldWeight in 5135, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5135)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    22. 3.2008 16:18:27
  4. Croft, W.B.; Metzler, D.; Strohman, T.: Search engines : information retrieval in practice (2010) 0.00
    0.0017656671 = product of:
      0.0070626684 = sum of:
        0.0070626684 = product of:
          0.028250674 = sum of:
            0.028250674 = weight(_text_:based in 2605) [ClassicSimilarity], result of:
              0.028250674 = score(doc=2605,freq=2.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.19973516 = fieldWeight in 2605, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2605)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    For introductory information retrieval courses at the undergraduate and graduate level in computer science, information science and computer engineering departments. Written by a leader in the field of information retrieval, Search Engines: Information Retrieval in Practice, is designed to give undergraduate students the understanding and tools they need to evaluate, compare and modify search engines. Coverage of the underlying IR and mathematical models reinforce key concepts. The book's numerous programming exercises make extensive use of Galago, a Java-based open source search engine. SUPPLEMENTS / Extensive lecture slides (in PDF and PPT format) / Solutions to selected end of chapter problems (Instructors only) / Test collections for exercises / Galago search engine
  5. Bizer, C.; Heath, T.: Linked Data : evolving the web into a global data space (2011) 0.00
    0.0016646868 = product of:
      0.0066587473 = sum of:
        0.0066587473 = product of:
          0.02663499 = sum of:
            0.02663499 = weight(_text_:based in 4725) [ClassicSimilarity], result of:
              0.02663499 = score(doc=4725,freq=4.0), product of:
                0.14144066 = queryWeight, product of:
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.04694356 = queryNorm
                0.18831211 = fieldWeight in 4725, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0129938 = idf(docFreq=5906, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4725)
          0.25 = coord(1/4)
      0.25 = coord(1/4)
    
    Abstract
    The World Wide Web has enabled the creation of a global information space comprising linked documents. As the Web becomes ever more enmeshed with our daily lives, there is a growing desire for direct access to raw data not currently available on the Web or bound up in hypertext documents. Linked Data provides a publishing paradigm in which not only documents, but also data, can be a first class citizen of the Web, thereby enabling the extension of the Web with a global data space based on open standards - the Web of Data. In this Synthesis lecture we provide readers with a detailed technical introduction to Linked Data. We begin by outlining the basic principles of Linked Data, including coverage of relevant aspects of Web architecture. The remainder of the text is based around two main themes - the publication and consumption of Linked Data. Drawing on a practical Linked Data scenario, we provide guidance and best practices on: architectural approaches to publishing Linked Data; choosing URIs and vocabularies to identify and describe resources; deciding what data to return in a description of a resource on the Web; methods and frameworks for automated linking of data sets; and testing and debugging approaches for Linked Data deployments. We give an overview of existing Linked Data applications and then examine the architectures that are used to consume Linked Data from the Web, alongside existing tools and frameworks that enable these. Readers can expect to gain a rich technical understanding of Linked Data fundamentals, as the basis for application development, research or further study.