Search (7 results, page 1 of 1)

  • × classification_ss:"ST 306"
  1. Manning, C.D.; Raghavan, P.; Schütze, H.: Introduction to information retrieval (2008) 0.01
    0.014313856 = product of:
      0.057255425 = sum of:
        0.050225936 = weight(_text_:retrieval in 4041) [ClassicSimilarity], result of:
          0.050225936 = score(doc=4041,freq=30.0), product of:
            0.09700725 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.032069415 = queryNorm
            0.51775444 = fieldWeight in 4041, product of:
              5.477226 = tf(freq=30.0), with freq of:
                30.0 = termFreq=30.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=4041)
        0.0070294905 = product of:
          0.014058981 = sum of:
            0.014058981 = weight(_text_:system in 4041) [ClassicSimilarity], result of:
              0.014058981 = score(doc=4041,freq=2.0), product of:
                0.10100432 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.032069415 = queryNorm
                0.13919188 = fieldWeight in 4041, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4041)
          0.5 = coord(1/2)
      0.25 = coord(2/8)
    
    Abstract
    Class-tested and coherent, this textbook teaches information retrieval, including web search, text classification, and text clustering from basic concepts. Ideas are explained using examples and figures, making it perfect for introductory courses in information retrieval for advanced undergraduates and graduate students. Slides and additional exercises are available for lecturers. - This book provides what Salton and Van Rijsbergen both failed to achieve. Even more important, unlike some other books in IR, the authors appear to care about making the theory as accessible as possible to the reader, on occasion including short primers to certain topics or choosing to explain difficult concepts using simplified approaches. Its coverage [is] excellent, the quality of writing high and I was surprised how much I learned from reading it. I think the online resources are impressive.
    Content
    Inhalt: Boolean retrieval - The term vocabulary & postings lists - Dictionaries and tolerant retrieval - Index construction - Index compression - Scoring, term weighting & the vector space model - Computing scores in a complete search system - Evaluation in information retrieval - Relevance feedback & query expansion - XML retrieval - Probabilistic information retrieval - Language models for information retrieval - Text classification & Naive Bayes - Vector space classification - Support vector machines & machine learning on documents - Flat clustering - Hierarchical clustering - Matrix decompositions & latent semantic indexing - Web search basics - Web crawling and indexes - Link analysis Vgl. die digitale Fassung unter: http://nlp.stanford.edu/IR-book/pdf/irbookprint.pdf.
    LCSH
    Information retrieval
    RSWK
    Dokumentverarbeitung / Information Retrieval / Abfrageverarbeitung (GBV)
    Information Retrieval / Einführung (BVB)
    Subject
    Dokumentverarbeitung / Information Retrieval / Abfrageverarbeitung (GBV)
    Information Retrieval / Einführung (BVB)
    Information retrieval
  2. Semantic applications (2018) 0.01
    0.013033398 = product of:
      0.052133594 = sum of:
        0.03970709 = weight(_text_:retrieval in 5204) [ClassicSimilarity], result of:
          0.03970709 = score(doc=5204,freq=12.0), product of:
            0.09700725 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.032069415 = queryNorm
            0.40932083 = fieldWeight in 5204, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5204)
        0.012426502 = product of:
          0.024853004 = sum of:
            0.024853004 = weight(_text_:system in 5204) [ClassicSimilarity], result of:
              0.024853004 = score(doc=5204,freq=4.0), product of:
                0.10100432 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.032069415 = queryNorm
                0.24605882 = fieldWeight in 5204, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5204)
          0.5 = coord(1/2)
      0.25 = coord(2/8)
    
    LCSH
    Information storage and retrieval
    Information Storage and Retrieval
    RSWK
    Wissensbasiertes System
    Information Retrieval
    Subject
    Wissensbasiertes System
    Information Retrieval
    Information storage and retrieval
    Information Storage and Retrieval
  3. Multi-source, multilingual information extraction and summarization (2013) 0.01
    0.009603616 = product of:
      0.038414463 = sum of:
        0.012426502 = product of:
          0.024853004 = sum of:
            0.024853004 = weight(_text_:system in 978) [ClassicSimilarity], result of:
              0.024853004 = score(doc=978,freq=4.0), product of:
                0.10100432 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.032069415 = queryNorm
                0.24605882 = fieldWeight in 978, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=978)
          0.5 = coord(1/2)
        0.02598796 = product of:
          0.05197592 = sum of:
            0.05197592 = weight(_text_:etc in 978) [ClassicSimilarity], result of:
              0.05197592 = score(doc=978,freq=2.0), product of:
                0.17370372 = queryWeight, product of:
                  5.4164915 = idf(docFreq=533, maxDocs=44218)
                  0.032069415 = queryNorm
                0.2992217 = fieldWeight in 978, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.4164915 = idf(docFreq=533, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=978)
          0.5 = coord(1/2)
      0.25 = coord(2/8)
    
    Abstract
    Information extraction (IE) and text summarization (TS) are powerful technologies for finding relevant pieces of information in text and presenting them to the user in condensed form. The ongoing information explosion makes IE and TS critical for successful functioning within the information society. These technologies face particular challenges due to the inherent multi-source nature of the information explosion. The technologies must now handle not isolated texts or individual narratives, but rather large-scale repositories and streams---in general, in multiple languages---containing a multiplicity of perspectives, opinions, or commentaries on particular topics, entities or events. There is thus a need to adapt existing techniques and develop new ones to deal with these challenges. This volume contains a selection of papers that present a variety of methodologies for content identification and extraction, as well as for content fusion and regeneration. The chapters cover various aspects of the challenges, depending on the nature of the information sought---names vs. events,--- and the nature of the sources---news streams vs. image captions vs. scientific research papers, etc. This volume aims to offer a broad and representative sample of studies from this very active research field.
    RSWK
    Natürlichsprachiges System / Information Extraction / Automatische Inhaltsanalyse / Zusammenfassung / Aufsatzsammlung
    Subject
    Natürlichsprachiges System / Information Extraction / Automatische Inhaltsanalyse / Zusammenfassung / Aufsatzsammlung
  4. Helbig, H.: Knowledge representation and the semantics of natural language (2014) 0.00
    0.003248495 = product of:
      0.02598796 = sum of:
        0.02598796 = product of:
          0.05197592 = sum of:
            0.05197592 = weight(_text_:etc in 2396) [ClassicSimilarity], result of:
              0.05197592 = score(doc=2396,freq=2.0), product of:
                0.17370372 = queryWeight, product of:
                  5.4164915 = idf(docFreq=533, maxDocs=44218)
                  0.032069415 = queryNorm
                0.2992217 = fieldWeight in 2396, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.4164915 = idf(docFreq=533, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2396)
          0.5 = coord(1/2)
      0.125 = coord(1/8)
    
    Abstract
    Natural Language is not only the most important means of communication between human beings, it is also used over historical periods for the preservation of cultural achievements and their transmission from one generation to the other. During the last few decades, the flod of digitalized information has been growing tremendously. This tendency will continue with the globalisation of information societies and with the growing importance of national and international computer networks. This is one reason why the theoretical understanding and the automated treatment of communication processes based on natural language have such a decisive social and economic impact. In this context, the semantic representation of knowledge originally formulated in natural language plays a central part, because it connects all components of natural language processing systems, be they the automatic understanding of natural language (analysis), the rational reasoning over knowledge bases, or the generation of natural language expressions from formal representations. This book presents a method for the semantic representation of natural language expressions (texts, sentences, phrases, etc.) which can be used as a universal knowledge representation paradigm in the human sciences, like linguistics, cognitive psychology, or philosophy of language, as well as in computational linguistics and in artificial intelligence. It is also an attempt to close the gap between these disciplines, which to a large extent are still working separately.
  5. Manning, C.D.; Schütze, H.: Foundations of statistical natural language processing (2000) 0.00
    0.0024315526 = product of:
      0.019452421 = sum of:
        0.019452421 = weight(_text_:retrieval in 1603) [ClassicSimilarity], result of:
          0.019452421 = score(doc=1603,freq=2.0), product of:
            0.09700725 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.032069415 = queryNorm
            0.20052543 = fieldWeight in 1603, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=1603)
      0.125 = coord(1/8)
    
    Abstract
    Statistical approaches to processing natural language text have become dominant in recent years. This foundational text is the first comprehensive introduction to statistical Natural Language Processing (NLP) to appear. The book contains all the theory and algorithms needed for building NLP tools. It provides broad but rigorous coverage of mathematical and linguistic foundations, as well as detailed discussion of statistical methods, allowing students and researchers to construct their own implementations. The book covers collocation finding, word sense disambiguation, probabilistic parsing, information retrieval, and other applications.
  6. Helbig, H.: Wissensverarbeitung und die Semantik der natürlichen Sprache : Wissensrepräsentation mit MultiNet (2008) 0.00
    0.0010983578 = product of:
      0.008786863 = sum of:
        0.008786863 = product of:
          0.017573725 = sum of:
            0.017573725 = weight(_text_:system in 2731) [ClassicSimilarity], result of:
              0.017573725 = score(doc=2731,freq=2.0), product of:
                0.10100432 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.032069415 = queryNorm
                0.17398985 = fieldWeight in 2731, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2731)
          0.5 = coord(1/2)
      0.125 = coord(1/8)
    
    Abstract
    Das Buch gibt eine umfassende Darstellung einer Methodik zur Interpretation und Bedeutungsrepräsentation natürlichsprachlicher Ausdrücke. Diese Methodik der "Mehrschichtigen Erweiterten Semantischen Netze", das sogenannte MultiNet-Paradigma, ist sowohl für theoretische Untersuchungen als auch für die automatische Verarbeitung natürlicher Sprache auf dem Rechner geeignet. Im ersten Teil des zweiteiligen Buches werden grundlegende Probleme der semantischen Repräsentation von Wissen bzw. der semantischen Interpretation natürlichsprachlicher Phänomene behandelt. Der zweite Teil enthält eine systematische Zusammenstellung des gesamten Repertoires von Darstellungsmitteln, die jeweils nach einem einheitlichen Schema beschrieben werden. Er dient als Kompendium der im Buch verwendeten formalen Beschreibungsmittel von MultiNet. Die vorgestellten Ergebnisse sind eingebettet in ein System von Software-Werkzeugen, die eine praktische Nutzung der MultiNet-Darstellungsmittel als Formalismus zur Bedeutungsrepräsentation im Rahmen der automatischen Sprachverarbeitung sichern. Hierzu gehören: eine Werkbank für den Wissensingenieur, ein Übersetzungssystem zur automatischen Gewinnung von Bedeutungsdarstellungen natürlichsprachlicher Sätze und eine Werkbank für den Computerlexikographen. Der Inhalt des Buches beruht auf jahrzehntelanger Forschung auf dem Gebiet der automatischen Sprachverarbeitung und wurde mit Vorlesungen zur Künstlichen Intelligenz und Wissensverarbeitung an der TU Dresden und der FernUniversität Hagen wiederholt in der Hochschullehre eingesetzt. Als Vorkenntnisse werden beim Leser lediglich Grundlagen der traditionellen Grammatik und elementare Kenntnisse der Prädikatenlogik vorausgesetzt.
  7. Semantische Technologien : Grundlagen - Konzepte - Anwendungen (2012) 0.00
    0.0010873189 = product of:
      0.008698551 = sum of:
        0.008698551 = product of:
          0.017397102 = sum of:
            0.017397102 = weight(_text_:system in 167) [ClassicSimilarity], result of:
              0.017397102 = score(doc=167,freq=4.0), product of:
                0.10100432 = queryWeight, product of:
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.032069415 = queryNorm
                0.17224117 = fieldWeight in 167, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.1495528 = idf(docFreq=5152, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=167)
          0.5 = coord(1/2)
      0.125 = coord(1/8)
    
    RSWK
    Semantic Web / Information Extraction / Suche / Wissensbasiertes System / Aufsatzsammlung
    Subject
    Semantic Web / Information Extraction / Suche / Wissensbasiertes System / Aufsatzsammlung

Languages

Types

Subjects