Search (146 results, page 2 of 8)

  • × theme_ss:"Computerlinguistik"
  1. Proszeky, G.: Language technology tools in the translator's practice (1999) 0.00
    0.0024581011 = product of:
      0.01966481 = sum of:
        0.01966481 = product of:
          0.058994424 = sum of:
            0.058994424 = weight(_text_:29 in 6873) [ClassicSimilarity], result of:
              0.058994424 = score(doc=6873,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.5441145 = fieldWeight in 6873, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6873)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    30. 3.2002 18:29:40
  2. McMahon, J.G.; Smith, F.J.: Improved statistical language model performance with automatic generated word hierarchies (1996) 0.00
    0.002435989 = product of:
      0.019487912 = sum of:
        0.019487912 = product of:
          0.058463734 = sum of:
            0.058463734 = weight(_text_:22 in 3164) [ClassicSimilarity], result of:
              0.058463734 = score(doc=3164,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.5416616 = fieldWeight in 3164, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3164)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Source
    Computational linguistics. 22(1996) no.2, S.217-248
  3. Ruge, G.: ¬A spreading activation network for automatic generation of thesaurus relationships (1991) 0.00
    0.002435989 = product of:
      0.019487912 = sum of:
        0.019487912 = product of:
          0.058463734 = sum of:
            0.058463734 = weight(_text_:22 in 4506) [ClassicSimilarity], result of:
              0.058463734 = score(doc=4506,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.5416616 = fieldWeight in 4506, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4506)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    8.10.2000 11:52:22
  4. Somers, H.: Example-based machine translation : Review article (1999) 0.00
    0.002435989 = product of:
      0.019487912 = sum of:
        0.019487912 = product of:
          0.058463734 = sum of:
            0.058463734 = weight(_text_:22 in 6672) [ClassicSimilarity], result of:
              0.058463734 = score(doc=6672,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.5416616 = fieldWeight in 6672, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6672)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    31. 7.1996 9:22:19
  5. New tools for human translators (1997) 0.00
    0.002435989 = product of:
      0.019487912 = sum of:
        0.019487912 = product of:
          0.058463734 = sum of:
            0.058463734 = weight(_text_:22 in 1179) [ClassicSimilarity], result of:
              0.058463734 = score(doc=1179,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.5416616 = fieldWeight in 1179, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1179)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    31. 7.1996 9:22:19
  6. Baayen, R.H.; Lieber, H.: Word frequency distributions and lexical semantics (1997) 0.00
    0.002435989 = product of:
      0.019487912 = sum of:
        0.019487912 = product of:
          0.058463734 = sum of:
            0.058463734 = weight(_text_:22 in 3117) [ClassicSimilarity], result of:
              0.058463734 = score(doc=3117,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.5416616 = fieldWeight in 3117, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3117)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    28. 2.1999 10:48:22
  7. ¬Der Student aus dem Computer (2023) 0.00
    0.002435989 = product of:
      0.019487912 = sum of:
        0.019487912 = product of:
          0.058463734 = sum of:
            0.058463734 = weight(_text_:22 in 1079) [ClassicSimilarity], result of:
              0.058463734 = score(doc=1079,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.5416616 = fieldWeight in 1079, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1079)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    27. 1.2023 16:22:55
  8. Fang, L.; Tuan, L.A.; Hui, S.C.; Wu, L.: Syntactic based approach for grammar question retrieval (2018) 0.00
    0.0022137975 = product of:
      0.01771038 = sum of:
        0.01771038 = product of:
          0.05313114 = sum of:
            0.05313114 = weight(_text_:problem in 5086) [ClassicSimilarity], result of:
              0.05313114 = score(doc=5086,freq=6.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.4061259 = fieldWeight in 5086, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5086)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    With the popularity of online educational platforms, English learners can learn and practice no matter where they are and what they do. English grammar is one of the important components in learning English. To learn English grammar effectively, it requires students to practice questions containing focused grammar knowledge. In this paper, we study a novel problem of retrieving English grammar questions with similar grammatical focus. Since the grammatical focus similarity is different from textual similarity or sentence syntactic similarity, existing approaches cannot be applied directly to our problem. To address this problem, we propose a syntactic based approach for English grammar question retrieval which can retrieve related grammar questions with similar grammatical focus effectively. In the proposed syntactic based approach, we first propose a new syntactic tree, namely parse-key tree, to capture English grammar questions' grammatical focus. Next, we propose two kernel functions, namely relaxed tree kernel and part-of-speech order kernel, to compute the similarity between two parse-key trees of the query and grammar questions in the collection. Then, the retrieved grammar questions are ranked according to the similarity between the parse-key trees. In addition, if a query is submitted together with answer choices, conceptual similarity and textual similarity are also incorporated to further improve the retrieval accuracy. The performance results have shown that our proposed approach outperforms the state-of-the-art methods based on statistical analysis and syntactic analysis.
  9. Rahmstorf, G.: Information retrieval using conceptual representations of phrases (1994) 0.00
    0.0021690698 = product of:
      0.017352559 = sum of:
        0.017352559 = product of:
          0.052057672 = sum of:
            0.052057672 = weight(_text_:problem in 7862) [ClassicSimilarity], result of:
              0.052057672 = score(doc=7862,freq=4.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.39792046 = fieldWeight in 7862, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.046875 = fieldNorm(doc=7862)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    The information retrieval problem is described starting from an analysis of the concepts 'user's information request' and 'information offerings of texts'. It is shown that natural language phrases are a more adequate medium for expressing information requests and information offerings than character string based query and indexing languages complemented by Boolean oprators. The phrases must be represented as concepts to reach a language invariant level for rule based relevance analysis. The special type of representation called advanced thesaurus is used for the semantic representation of natural language phrases and for relevance processing. The analysis of the retrieval problem leads to a symmetric system structure
  10. Zue, V.: Dein Computer spricht mir dir (1999) 0.00
    0.0021069439 = product of:
      0.01685555 = sum of:
        0.01685555 = product of:
          0.05056665 = sum of:
            0.05056665 = weight(_text_:29 in 4244) [ClassicSimilarity], result of:
              0.05056665 = score(doc=4244,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.46638384 = fieldWeight in 4244, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4244)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    31.12.1996 19:29:41
  11. Mustafa El Hadi, W.: Evaluating human language technology : general applications to information access and management (2002) 0.00
    0.0021069439 = product of:
      0.01685555 = sum of:
        0.01685555 = product of:
          0.05056665 = sum of:
            0.05056665 = weight(_text_:29 in 1840) [ClassicSimilarity], result of:
              0.05056665 = score(doc=1840,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.46638384 = fieldWeight in 1840, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=1840)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Source
    Knowledge organization. 29(2002) nos.3/4, S.124-134
  12. Müller, P.: Text-Automat mit Tücken (2023) 0.00
    0.0021069439 = product of:
      0.01685555 = sum of:
        0.01685555 = product of:
          0.05056665 = sum of:
            0.05056665 = weight(_text_:29 in 481) [ClassicSimilarity], result of:
              0.05056665 = score(doc=481,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.46638384 = fieldWeight in 481, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=481)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Source
    Pirmasenser Zeitung. Nr. 29 vom 03.02.2023, S.2
  13. Byrne, C.C.; McCracken, S.A.: ¬An adaptive thesaurus employing semantic distance, relational inheritance and nominal compound interpretation for linguistic support of information retrieval (1999) 0.00
    0.0020879905 = product of:
      0.016703924 = sum of:
        0.016703924 = product of:
          0.05011177 = sum of:
            0.05011177 = weight(_text_:22 in 4483) [ClassicSimilarity], result of:
              0.05011177 = score(doc=4483,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.46428138 = fieldWeight in 4483, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4483)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    15. 3.2000 10:22:37
  14. Boleda, G.; Evert, S.: Multiword expressions : a pain in the neck of lexical semantics (2009) 0.00
    0.0020879905 = product of:
      0.016703924 = sum of:
        0.016703924 = product of:
          0.05011177 = sum of:
            0.05011177 = weight(_text_:22 in 4888) [ClassicSimilarity], result of:
              0.05011177 = score(doc=4888,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.46428138 = fieldWeight in 4888, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4888)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    1. 3.2013 14:56:22
  15. Monnerjahn, P.: Vorsprung ohne Technik : Übersetzen: Computer und Qualität (2000) 0.00
    0.0020879905 = product of:
      0.016703924 = sum of:
        0.016703924 = product of:
          0.05011177 = sum of:
            0.05011177 = weight(_text_:22 in 5429) [ClassicSimilarity], result of:
              0.05011177 = score(doc=5429,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.46428138 = fieldWeight in 5429, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=5429)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Source
    c't. 2000, H.22, S.230-231
  16. Stede, M.: Lexicalization in natural language generation : a survey (1994/95) 0.00
    0.0020450184 = product of:
      0.016360147 = sum of:
        0.016360147 = product of:
          0.04908044 = sum of:
            0.04908044 = weight(_text_:problem in 1913) [ClassicSimilarity], result of:
              0.04908044 = score(doc=1913,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.375163 = fieldWeight in 1913, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1913)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    In natural language generation, a meaning representation of some kind is successively transformed into a sentence or a text. Naturally, a central subtask of this problem is the choice of words, or lexicalization. Proposes 4 major issues that determine how a generator tackles lexicalization, and surveys the contributions that research have made to them. Identifies open problems, and sketches a possible direction for research
  17. Kurz, C.: Womit sich Strafverfolger bald befassen müssen : ChatGPT (2023) 0.00
    0.0020450184 = product of:
      0.016360147 = sum of:
        0.016360147 = product of:
          0.04908044 = sum of:
            0.04908044 = weight(_text_:problem in 203) [ClassicSimilarity], result of:
              0.04908044 = score(doc=203,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.375163 = fieldWeight in 203, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0625 = fieldNorm(doc=203)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    Ein Europol-Bericht widmet sich den Folgen von ChatGPT, wenn Kriminelle die Fähigkeiten des Chatbots für sich ausnutzen: Es drohe vermehrt Phishing und noch mehr Desinformation. Ein Problem für die Strafverfolgung könne auch automatisiert erzeugter bösartiger Quellcode sein.
  18. Griffiths, T.L.; Steyvers, M.: ¬A probabilistic approach to semantic representation (2002) 0.00
    0.0019864459 = product of:
      0.015891567 = sum of:
        0.015891567 = product of:
          0.047674697 = sum of:
            0.047674697 = weight(_text_:29 in 3671) [ClassicSimilarity], result of:
              0.047674697 = score(doc=3671,freq=4.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.43971092 = fieldWeight in 3671, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3671)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    29. 6.2015 14:55:01
    29. 6.2015 16:09:05
  19. Ahlgren, P.; Kekäläinen, J.: Indexing strategies for Swedish full text retrieval under different user scenarios (2007) 0.00
    0.0018075579 = product of:
      0.014460463 = sum of:
        0.014460463 = product of:
          0.04338139 = sum of:
            0.04338139 = weight(_text_:problem in 896) [ClassicSimilarity], result of:
              0.04338139 = score(doc=896,freq=4.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.33160037 = fieldWeight in 896, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=896)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    This paper deals with Swedish full text retrieval and the problem of morphological variation of query terms in the document database. The effects of combination of indexing strategies with query terms on retrieval effectiveness were studied. Three of five tested combinations involved indexing strategies that used conflation, in the form of normalization. Further, two of these three combinations used indexing strategies that employed compound splitting. Normalization and compound splitting were performed by SWETWOL, a morphological analyzer for the Swedish language. A fourth combination attempted to group related terms by right hand truncation of query terms. The four combinations were compared to each other and to a baseline combination, where no attempt was made to counteract the problem of morphological variation of query terms in the document database. The five combinations were evaluated under six different user scenarios, where each scenario simulated a certain user type. The four alternative combinations outperformed the baseline, for each user scenario. The truncation combination had the best performance under each user scenario. The main conclusion of the paper is that normalization and right hand truncation (performed by a search expert) enhanced retrieval effectiveness in comparison to the baseline. The performance of the three combinations of indexing strategies with query terms based on normalization was not far below the performance of the truncation combination.
  20. Hodgson, J.P.E.: Knowledge representation and language in AI (1991) 0.00
    0.0018075579 = product of:
      0.014460463 = sum of:
        0.014460463 = product of:
          0.04338139 = sum of:
            0.04338139 = weight(_text_:problem in 1529) [ClassicSimilarity], result of:
              0.04338139 = score(doc=1529,freq=4.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.33160037 = fieldWeight in 1529, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1529)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    The aim of this book is to highlight the relationship between knowledge representation and language in artificial intelligence, and in particular on the way in which the choice of representation influences the language used to discuss a problem - and vice versa. Opening with a discussion of knowledge representation methods, and following this with a look at reasoning methods, the author begins to make his case for the intimate relationship between language and representation. He shows how each representation method fits particularly well with some reasoning methods and less so with others, using specific languages as examples. The question of representation change, an important and complex issue about which very little is known, is addressed. Dr Hodgson gathers together recent work on problem solving, showing how, in some cases, it has been possible to use representation changes to recast problems into a language that makes them easier to solve. The author maintains throughout that the relationships that this book explores lie at the heart of the construction of large systems, examining a number of the current large AI systems from the viewpoint of representation and language to prove his point.

Years

Languages

  • e 102
  • d 39
  • ru 3
  • More… Less…

Types

  • a 123
  • el 17
  • m 10
  • s 4
  • x 4
  • p 2
  • d 1
  • More… Less…

Classifications