Search (3 results, page 1 of 1)

  • × classification_ss:"BFP (FH K)"
  1. Jacquemin, C.: Spotting and discovering terms through natural language processing (2001) 0.02
    0.024606533 = product of:
      0.049213067 = sum of:
        0.049213067 = product of:
          0.09842613 = sum of:
            0.09842613 = weight(_text_:language in 119) [ClassicSimilarity], result of:
              0.09842613 = score(doc=119,freq=10.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.48463053 = fieldWeight in 119, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=119)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    In this book Christian Jacquemin shows how the power of natural language processing (NLP) can be used to advance text indexing and information retrieval (IR). Jacquemin's novel tool is FASTR, a parser that normalizes terms and recognizes term variants. Since there are more meanings in a language than there are words, FASTR uses a metagrammar composed of shallow linguistic transformations that describe the morphological, syntactic, semantic, and pragmatic variations of words and terms. The acquired parsed terms can then be applied for precise retrieval and assembly of information. The use of a corpus-based unification grammar to define, recognize, and combine term variants from their base forms allows for intelligent information access to, or "linguistic data tuning" of, heterogeneous texts. FASTR can be used to do automatic controlled indexing, to carry out content-based Web searches through conceptually related alternative query formulations, to abstract scientific and technical extracts, and even to translate and collect terms from multilingual material. Jacquemin provides a comprehensive account of the method and implementation of this innovative retrieval technique for text processing.
    LCSH
    Language and languages / Variation / Data processing
    Subject
    Language and languages / Variation / Data processing
  2. Ramisch, C.: Multiword expressions acquisition : a generic and open framework (2015) 0.02
    0.019685227 = product of:
      0.039370455 = sum of:
        0.039370455 = product of:
          0.07874091 = sum of:
            0.07874091 = weight(_text_:language in 1649) [ClassicSimilarity], result of:
              0.07874091 = score(doc=1649,freq=10.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.38770443 = fieldWeight in 1649, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1649)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This book is an excellent introduction to multiword expressions. It provides a unique, comprehensive and up-to-date overview of this exciting topic in computational linguistics. The first part describes the diversity and richness of multiword expressions, including many examples in several languages. These constructions are not only complex and arbitrary, but also much more frequent than one would guess, making them a real nightmare for natural language processing applications. The second part introduces a new generic framework for automatic acquisition of multiword expressions from texts. Furthermore, it describes the accompanying free software tool, the mwetoolkit, which comes in handy when looking for expressions in texts (regardless of the language). Evaluation is greatly emphasized, underlining the fact that results depend on parameters like corpus size, language, MWE type, etc. The last part contains solid experimental results and evaluates the mwetoolkit, demonstrating its usefulness for computer-assisted lexicography and machine translation. This is the first book to cover the whole pipeline of multiword expression acquisition in a single volume. It is addresses the needs of students and researchers in computational and theoretical linguistics, cognitive sciences, artificial intelligence and computer science. Its good balance between computational and linguistic views make it the perfect starting point for anyone interested in multiword expressions, language and text processing in general.
    Series
    Theory and applications of natural language processing
  3. Bowker, L.; Ciro, J.B.: Machine translation and global research : towards improved machine translation literacy in the scholarly community (2019) 0.01
    0.008803501 = product of:
      0.017607002 = sum of:
        0.017607002 = product of:
          0.035214003 = sum of:
            0.035214003 = weight(_text_:language in 5970) [ClassicSimilarity], result of:
              0.035214003 = score(doc=5970,freq=2.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.17338668 = fieldWeight in 5970, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5970)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    In the global research community, English has become the main language of scholarly publishing in many disciplines. At the same time, online machine translation systems have become increasingly easy to access and use. Is this a researcher's match made in heaven, or the road to publication perdition? Here Lynne Bowker and Jairo Buitrago Ciro introduce the concept of machine translation literacy, a new kind of literacy for scholars and librarians in the digital age. For scholars, they explain how machine translation works, how it is (or could be) used for scholarly communication, and how both native and non-native English-speakers can write in a translation-friendly way in order to harness its potential. Native English speakers can continue to write in English, but expand the global reach of their research by making it easier for their peers around the world to access and understand their works, while non-native English speakers can write in their mother tongues, but leverage machine translation technology to help them produce draft publications in English. For academic librarians, the authors provide a framework for supporting researchers in all disciplines as they grapple with producing translation-friendly texts and using machine translation for scholarly communication - a form of support that will only become more important as campuses become increasingly international and as universities continue to strive to excel on the global stage. Machine Translation and Global Research is a must-read for scientists, researchers, students, and librarians eager to maximize the global reach and impact of any form of scholarly work.