Search (6 results, page 1 of 1)

  • × classification_ss:"ST 205"
  1. Handbuch Internet-Suchmaschinen 2 : Neue Entwicklungen in der Web-Suche (2011) 0.02
    0.019630127 = product of:
      0.05889038 = sum of:
        0.05889038 = weight(_text_:query in 522) [ClassicSimilarity], result of:
          0.05889038 = score(doc=522,freq=2.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.25674784 = fieldWeight in 522, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.0390625 = fieldNorm(doc=522)
      0.33333334 = coord(1/3)
    
    Content
    Titel der einzelnen Kapitel: Social Search Folksonomies und Kollaborative Informationsdienste: Eine Alternative zur Websuche? Query Understanding Semantic Search Suche in Multimediaarchiven und Kultureinrichtungen Wissenschaftliche Suchmaschinen Journalistische Recherche im Internet Evaluierung von Suchmaschinen Usability und User Experience in Suchmaschinen Search Engine Bias Web Monitoring Tools für das Monitoring
  2. Manning, C.D.; Raghavan, P.; Schütze, H.: Introduction to information retrieval (2008) 0.02
    0.015704103 = product of:
      0.047112305 = sum of:
        0.047112305 = weight(_text_:query in 4041) [ClassicSimilarity], result of:
          0.047112305 = score(doc=4041,freq=2.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.20539828 = fieldWeight in 4041, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.03125 = fieldNorm(doc=4041)
      0.33333334 = coord(1/3)
    
    Content
    Inhalt: Boolean retrieval - The term vocabulary & postings lists - Dictionaries and tolerant retrieval - Index construction - Index compression - Scoring, term weighting & the vector space model - Computing scores in a complete search system - Evaluation in information retrieval - Relevance feedback & query expansion - XML retrieval - Probabilistic information retrieval - Language models for information retrieval - Text classification & Naive Bayes - Vector space classification - Support vector machines & machine learning on documents - Flat clustering - Hierarchical clustering - Matrix decompositions & latent semantic indexing - Web search basics - Web crawling and indexes - Link analysis Vgl. die digitale Fassung unter: http://nlp.stanford.edu/IR-book/pdf/irbookprint.pdf.
  3. Hitzler, P.; Krötzsch, M.; Rudolph, S.: Foundations of Semantic Web technologies (2010) 0.02
    0.015704103 = product of:
      0.047112305 = sum of:
        0.047112305 = weight(_text_:query in 359) [ClassicSimilarity], result of:
          0.047112305 = score(doc=359,freq=2.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.20539828 = fieldWeight in 359, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.03125 = fieldNorm(doc=359)
      0.33333334 = coord(1/3)
    
    Abstract
    This text introduces the standardized knowledge representation languages for modeling ontologies operating at the core of the semantic web. It covers RDF schema, Web Ontology Language (OWL), rules, query languages, the OWL 2 revision, and the forthcoming Rule Interchange Format (RIF). A 2010 CHOICE Outstanding Academic Title ! The nine chapters of the book guide the reader through the major foundational languages for the semantic Web and highlight the formal semantics. ! the book has very interesting supporting material and exercises, is oriented to W3C standards, and provides the necessary foundations for the semantic Web. It will be easy to follow by the computer scientist who already has a basic background on semantic Web issues; it will also be helpful for both self-study and teaching purposes. I recommend this book primarily as a complementary textbook for a graduate or undergraduate course in a computer science or a Web science academic program. --Computing Reviews, February 2010 This book is unique in several respects. It contains an in-depth treatment of all the major foundational languages for the Semantic Web and provides a full treatment of the underlying formal semantics, which is central to the Semantic Web effort. It is also the very first textbook that addresses the forthcoming W3C recommended standards OWL 2 and RIF. Furthermore, the covered topics and underlying concepts are easily accessible for the reader due to a clear separation of syntax and semantics ! I am confident this book will be well received and play an important role in training a larger number of students who will seek to become proficient in this growing discipline.
  4. Spink, A.; Jansen, B.J.: Web searching : public searching of the Web (2004) 0.01
    0.009815063 = product of:
      0.02944519 = sum of:
        0.02944519 = weight(_text_:query in 1443) [ClassicSimilarity], result of:
          0.02944519 = score(doc=1443,freq=2.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.12837392 = fieldWeight in 1443, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.01953125 = fieldNorm(doc=1443)
      0.33333334 = coord(1/3)
    
    Footnote
    Rez. in: Information - Wissenschaft und Praxis 56(2004) H.1, S.61-62 (D. Lewandowski): "Die Autoren des vorliegenden Bandes haben sich in den letzten Jahren durch ihre zahlreichen Veröffentlichungen zum Verhalten von Suchmaschinen-Nutzern einen guten Namen gemacht. Das nun erschienene Buch bietet eine Zusammenfassung der verstreut publizierten Aufsätze und stellt deren Ergebnisse in den Kontext eines umfassenderen Forschungsansatzes. Spink und Jansen verwenden zur Analyse des Nutzungsverhaltens query logs von Suchmaschinen. In diesen werden vom Server Informationen protokolliert, die die Anfragen an diesen Server betreffen. Daten, die aus diesen Dateien gewonnen werden können, sind unter anderem die gestellten Suchanfragen, die Adresse des Rechners, von dem aus die Anfrage gestellt wurde, sowie die aus den Trefferlisten ausgewählten Dokumente. Der klare Vorteil der Analyse von Logfiles liegt in der Möglichkeit, große Datenmengen ohne hohen personellen Aufwand erheben zu können. Die Daten einer Vielzahl anonymer Nutzer können analysiert werden; ohne dass dabei die Datenerhebung das Nutzerverhalten beeinflusst. Dies ist bei Suchmaschinen von besonderer Bedeutung, weil sie im Gegensatz zu den meisten anderen professionellen Information-Retrieval-Systemen nicht nur im beruflichen Kontext, sondern auch (und vor allem) privat genutzt werden. Das Bild des Nutzungsverhaltens wird in Umfragen und Laboruntersuchungen verfälscht, weil Nutzer ihr Anfrageverhalten falsch einschätzen oder aber die Themen ihrer Anfragen nicht nennen möchten. Hier ist vor allem an Suchanfragen, die auf medizinische oder pornographische Inhalte gerichtet sind, zu denken. Die Analyse von Logfiles ist allerdings auch mit Problemen behaftet: So sind nicht alle gewünschten Daten überhaupt in den Logfiles enthalten (es fehlen alle Informationen über den einzelnen Nutzer), es werden keine qualitativen Informationen wie etwa der Grund einer Suche erfasst und die Logfiles sind aufgrund technischer Gegebenheiten teils unvollständig. Die Autoren schließen aus den genannten Vor- und Nachteilen, dass sich Logfiles gut für die Auswertung des Nutzerverhaltens eignen, bei der Auswertung jedoch die Ergebnisse von Untersuchungen, welche andere Methoden verwenden, berücksichtigt werden sollten.
  5. Rogers, R.: Information politics on the Web (2004) 0.01
    0.007852051 = product of:
      0.023556152 = sum of:
        0.023556152 = weight(_text_:query in 442) [ClassicSimilarity], result of:
          0.023556152 = score(doc=442,freq=2.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.10269914 = fieldWeight in 442, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.015625 = fieldNorm(doc=442)
      0.33333334 = coord(1/3)
    
    Footnote
    Chapter 2 examines the politics of information retrieval in the context of collaborative filtering techniques. Rogers begins by discussing the underpinnings of modern search engine design by examining medieval practices of knowledge seeking, following up with a critique of the collaborative filtering techniques. Rogers's major contention is that collaborative filtering rids us of user idiosyncrasies as search query strings, preferences, and recommendations are shared among users and without much care for the differences among them, both in terms of their innate characteristics and also their search goals. To illustrate Rogers' critiques of collaborative filtering, he describes an information searching experiment that he conducted with students at University of Vienna and University of Amsterdam. Students were asked to search for information on Viagra. As one can imagine, depending on a number of issues, not the least of which is what sources did one extract information from, a student would find different accounts of reality about Viagra, everything from a medical drug to a black-market drug ideal for underground trade. Rogers described how information on the Web differed from official accounts for certain events. The information on the Web served as an alternative reality. Chapter 3 describes the Web as a dynamic debate-mapping tool, a political instrument. Rogers introduces the "Issue Barometer," an information instrument that measures the social pressure on a topic being debated by analyzing data available from the Web. Measures used by the Issue Barometer include temperature of the issue (cold to hot), activity level of the debate (mild to intense), and territorialization (one country to many countries). The Issues Barometer is applied to an illustrative case of the public debate surrounding food safety in the Netherlands in 2001. Chapter 4 introduces "The Web Issue Index," which provides an indication of leading societal issues discussed on the Web. The empirical research on the Web Issues Index was conducted on the Genoa G8 Summit in 1999 and the anti-globalization movement. Rogers focus here was to examine the changing nature of prominent issues over time, i.e., how issues gained and lost attention and traction over time.
  6. Rosenfeld, L.; Morville, P.: Information architecture for the World Wide Web : designing large-scale Web sites (2007) 0.01
    0.0055721086 = product of:
      0.016716326 = sum of:
        0.016716326 = product of:
          0.03343265 = sum of:
            0.03343265 = weight(_text_:22 in 5135) [ClassicSimilarity], result of:
              0.03343265 = score(doc=5135,freq=2.0), product of:
                0.1728227 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049352113 = queryNorm
                0.19345059 = fieldWeight in 5135, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5135)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 3.2008 16:18:27

Languages

Types