Search (26 results, page 1 of 2)

  • × language_ss:"d"
  • × theme_ss:"Computerlinguistik"
  • × year_i:[2010 TO 2020}
  1. Rötzer, F.: KI-Programm besser als Menschen im Verständnis natürlicher Sprache (2018) 0.01
    0.012431266 = product of:
      0.024862532 = sum of:
        0.024862532 = sum of:
          0.0024318986 = weight(_text_:a in 4217) [ClassicSimilarity], result of:
            0.0024318986 = score(doc=4217,freq=2.0), product of:
              0.04772363 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.041389145 = queryNorm
              0.050957955 = fieldWeight in 4217, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.03125 = fieldNorm(doc=4217)
          0.022430632 = weight(_text_:22 in 4217) [ClassicSimilarity], result of:
            0.022430632 = score(doc=4217,freq=2.0), product of:
              0.14493774 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041389145 = queryNorm
              0.15476047 = fieldWeight in 4217, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=4217)
      0.5 = coord(1/2)
    
    Date
    22. 1.2018 11:32:44
    Type
    a
  2. Lezius, W.: Morphy - Morphologie und Tagging für das Deutsche (2013) 0.01
    0.011215316 = product of:
      0.022430632 = sum of:
        0.022430632 = product of:
          0.044861265 = sum of:
            0.044861265 = weight(_text_:22 in 1490) [ClassicSimilarity], result of:
              0.044861265 = score(doc=1490,freq=2.0), product of:
                0.14493774 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.041389145 = queryNorm
                0.30952093 = fieldWeight in 1490, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1490)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 3.2015 9:30:24
  3. Biselli, A.: Unter Generalverdacht durch Algorithmen (2014) 0.00
    0.002579418 = product of:
      0.005158836 = sum of:
        0.005158836 = product of:
          0.010317672 = sum of:
            0.010317672 = weight(_text_:a in 809) [ClassicSimilarity], result of:
              0.010317672 = score(doc=809,freq=4.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.2161963 = fieldWeight in 809, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=809)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  4. Karlova-Bourbonus, N.: Automatic detection of contradictions in texts (2018) 0.00
    0.0018800576 = product of:
      0.0037601152 = sum of:
        0.0037601152 = product of:
          0.0075202305 = sum of:
            0.0075202305 = weight(_text_:a in 5976) [ClassicSimilarity], result of:
              0.0075202305 = score(doc=5976,freq=34.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.15757877 = fieldWeight in 5976, product of:
                  5.8309517 = tf(freq=34.0), with freq of:
                    34.0 = termFreq=34.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=5976)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Natural language contradictions are of complex nature. As will be shown in Chapter 5, the realization of contradictions is not limited to the examples such as Socrates is a man and Socrates is not a man (under the condition that Socrates refers to the same object in the real world), which is discussed by Aristotle (Section 3.1.1). Empirical evidence (see Chapter 5 for more details) shows that only a few contradictions occurring in the real life are of that explicit (prototypical) kind. Rather, con-tradictions make use of a variety of natural language devices such as, e.g., paraphrasing, synonyms and antonyms, passive and active voice, diversity of negation expression, and figurative linguistic means such as idioms, irony, and metaphors. Additionally, the most so-phisticated kind of contradictions, the so-called implicit contradictions, can be found only when applying world knowledge and after conducting a sequence of logical operations such as e.g. in: (1.1) The first prize was given to the experienced grandmaster L. Stein who, in total, col-lected ten points (7 wins and 3 draws). Those familiar with the chess rules know that a chess player gets one point for winning and zero points for losing the game. In case of a draw, each player gets a half point. Built on this idea and by conducting some simple mathematical operations, we can infer that in the case of 7 wins and 3 draws (the second part of the sentence), a player can only collect 8.5 points and not 10 points. Hence, we observe that there is a contradiction between the first and the second parts of the sentence.
    Implicit contradictions will only partially be the subject of the present study, aiming primarily at identifying the realization mechanism and cues (Chapter 5) as well as finding the parts of contradictions by applying the state of the art algorithms for natural language processing without conducting deep meaning processing. Further in focus are the explicit and implicit contradictions that can be detected by means of explicit linguistic, structural, lexical cues, and by conducting some additional processing operations (e.g., counting the sum in order to detect contradictions arising from numerical divergencies). One should note that an additional complexity in finding contradictions can arise in case parts of the contradictions occur on different levels of realization. Thus, a contradiction can be observed on the word- and phrase-level, such as in a married bachelor (for variations of contradictions on lexical level, see Ganeev 2004), on the sentence level - between parts of a sentence or between two or more sentences, or on the text level - between the portions of a text or between the whole texts such as a contradiction between the Bible and the Quran, for example. Only contradictions arising at the level of single sentences occurring in one or more texts, as well as parts of a sentence, will be considered for the purpose of this study. Though the focus of interest will be on single sentences, it will make use of text particularities such as coreference resolution without establishing the referents in the real world. Finally, another aspect to be considered is that parts of the contradictions are not neces-sarily to appear at the same time. They can be separated by many years and centuries with or without time expression making their recognition by human and detection by machine challenging. According to Aristotle's ontological version of the LNC (Section 3.1.1), how-ever, the same time reference is required in order for two statements to be judged as a contradiction. Taking this into account, we set the borders for the study by limiting the ana-lyzed textual data thematically (only nine world events) and temporally (three days after the reported event had happened) (Section 5.1). No sophisticated time processing will thus be conducted.
  5. Altmann, E.G.; Cristadoro, G.; Esposti, M.D.: On the origin of long-range correlations in texts (2012) 0.00
    0.001823924 = product of:
      0.003647848 = sum of:
        0.003647848 = product of:
          0.007295696 = sum of:
            0.007295696 = weight(_text_:a in 330) [ClassicSimilarity], result of:
              0.007295696 = score(doc=330,freq=8.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.15287387 = fieldWeight in 330, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=330)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The complexity of human interactions with social and natural phenomena is mirrored in the way we describe our experiences through natural language. In order to retain and convey such a high dimensional information, the statistical properties of our linguistic output has to be highly correlated in time. An example are the robust observations, still largely not understood, of correlations on arbitrary long scales in literary texts. In this paper we explain how long-range correlations flow from highly structured linguistic levels down to the building blocks of a text (words, letters, etc..). By combining calculations and data analysis we show that correlations take form of a bursty sequence of events once we approach the semantically relevant topics of the text. The mechanisms we identify are fairly general and can be equally applied to other hierarchical settings.
    Type
    a
  6. Engerer, V.: Indexierungstheorie für Linguisten : zu einigen natürlichsprachlichen Zügen in künstlichen Indexsprachen (2014) 0.00
    0.001823924 = product of:
      0.003647848 = sum of:
        0.003647848 = product of:
          0.007295696 = sum of:
            0.007295696 = weight(_text_:a in 3339) [ClassicSimilarity], result of:
              0.007295696 = score(doc=3339,freq=2.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.15287387 = fieldWeight in 3339, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3339)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  7. Voss, O.: Übersetzer überflüssig? : Sprachsoftware DeepL und Acrolinx (2019) 0.00
    0.001823924 = product of:
      0.003647848 = sum of:
        0.003647848 = product of:
          0.007295696 = sum of:
            0.007295696 = weight(_text_:a in 4981) [ClassicSimilarity], result of:
              0.007295696 = score(doc=4981,freq=2.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.15287387 = fieldWeight in 4981, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4981)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  8. Hahn, U.: Methodische Grundlagen der Informationslinguistik (2013) 0.00
    0.0015199365 = product of:
      0.003039873 = sum of:
        0.003039873 = product of:
          0.006079746 = sum of:
            0.006079746 = weight(_text_:a in 719) [ClassicSimilarity], result of:
              0.006079746 = score(doc=719,freq=2.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.12739488 = fieldWeight in 719, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=719)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  9. Engerer, V.: Informationswissenschaft und Linguistik. : kurze Geschichte eines fruchtbaren interdisziplinäaren Verhäaltnisses in drei Akten (2012) 0.00
    0.0015199365 = product of:
      0.003039873 = sum of:
        0.003039873 = product of:
          0.006079746 = sum of:
            0.006079746 = weight(_text_:a in 3376) [ClassicSimilarity], result of:
              0.006079746 = score(doc=3376,freq=2.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.12739488 = fieldWeight in 3376, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3376)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  10. Wolfangel, E.: Ich verstehe (2017) 0.00
    0.0015199365 = product of:
      0.003039873 = sum of:
        0.003039873 = product of:
          0.006079746 = sum of:
            0.006079746 = weight(_text_:a in 3976) [ClassicSimilarity], result of:
              0.006079746 = score(doc=3976,freq=2.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.12739488 = fieldWeight in 3976, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3976)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  11. Holland, M.: Erstes wissenschaftliches Buch eines Algorithmus' veröffentlicht (2019) 0.00
    0.0015046606 = product of:
      0.0030093212 = sum of:
        0.0030093212 = product of:
          0.0060186423 = sum of:
            0.0060186423 = weight(_text_:a in 5227) [ClassicSimilarity], result of:
              0.0060186423 = score(doc=5227,freq=4.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.12611452 = fieldWeight in 5227, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5227)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Der Wissenschaftsverlag Springer Nature hat nach eigenen Angaben das erste Buch veröffentlicht, das von einem Algorithmus verfasst wurde. Bei Springer Nature ist das nach Angaben des Wissenschaftsverlags erste maschinengenerierte Buch erschienen: "Lithium-Ion Batteries - A Machine-Generated Summary of Current Research" biete einen Überblick über die neuesten Forschungspublikationen über Lithium-Ionen-Batterien, erklärte die Goethe-Universität Frankfurt am Main. Dort wurde im Bereich Angewandte Computerlinguistik unter der Leitung von Christian Chiarcos jenes Verfahren entwickelt, das Textinhalte automatisch analysiert und relevante Publikationen auswählen kann. Es heißt "Beta Writer" und steht als Autor über dem Buch.
    Type
    a
  12. Heid, U.: Computerlinguistik zwischen Informationswissenschaft und multilingualer Kommunikation (2010) 0.00
    0.0012159493 = product of:
      0.0024318986 = sum of:
        0.0024318986 = product of:
          0.004863797 = sum of:
            0.004863797 = weight(_text_:a in 4018) [ClassicSimilarity], result of:
              0.004863797 = score(doc=4018,freq=2.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.10191591 = fieldWeight in 4018, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4018)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  13. Becks, D.; Schulz, J.M.: Domänenübergreifende Phrasenextraktion mithilfe einer lexikonunabhängigen Analysekomponente (2010) 0.00
    0.0012159493 = product of:
      0.0024318986 = sum of:
        0.0024318986 = product of:
          0.004863797 = sum of:
            0.004863797 = weight(_text_:a in 4661) [ClassicSimilarity], result of:
              0.004863797 = score(doc=4661,freq=2.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.10191591 = fieldWeight in 4661, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4661)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  14. Ludwig, B.; Reischer, J.: Informationslinguistik in Regensburg (2012) 0.00
    0.0012159493 = product of:
      0.0024318986 = sum of:
        0.0024318986 = product of:
          0.004863797 = sum of:
            0.004863797 = weight(_text_:a in 555) [ClassicSimilarity], result of:
              0.004863797 = score(doc=555,freq=2.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.10191591 = fieldWeight in 555, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=555)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  15. Baierer, K.; Zumstein, P.: Verbesserung der OCR in digitalen Sammlungen von Bibliotheken (2016) 0.00
    0.0012159493 = product of:
      0.0024318986 = sum of:
        0.0024318986 = product of:
          0.004863797 = sum of:
            0.004863797 = weight(_text_:a in 2818) [ClassicSimilarity], result of:
              0.004863797 = score(doc=2818,freq=2.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.10191591 = fieldWeight in 2818, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2818)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  16. Franke-Maier, M.: Computerlinguistik und Bibliotheken : Editorial (2016) 0.00
    0.0010747575 = product of:
      0.002149515 = sum of:
        0.002149515 = product of:
          0.00429903 = sum of:
            0.00429903 = weight(_text_:a in 3206) [ClassicSimilarity], result of:
              0.00429903 = score(doc=3206,freq=4.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.090081796 = fieldWeight in 3206, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3206)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Editor
    Ledl, A.
    Type
    a
  17. Strube, M.: Kreativ durch Analogien (2011) 0.00
    0.0010639556 = product of:
      0.0021279112 = sum of:
        0.0021279112 = product of:
          0.0042558224 = sum of:
            0.0042558224 = weight(_text_:a in 4805) [ClassicSimilarity], result of:
              0.0042558224 = score(doc=4805,freq=2.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.089176424 = fieldWeight in 4805, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4805)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  18. Rötzer, F.: Kann KI mit KI generierte Texte erkennen? (2019) 0.00
    0.0010639556 = product of:
      0.0021279112 = sum of:
        0.0021279112 = product of:
          0.0042558224 = sum of:
            0.0042558224 = weight(_text_:a in 3977) [ClassicSimilarity], result of:
              0.0042558224 = score(doc=3977,freq=2.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.089176424 = fieldWeight in 3977, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3977)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  19. Budin, G.: Zum Entwicklungsstand der Terminologiewissenschaft (2019) 0.00
    0.0010639556 = product of:
      0.0021279112 = sum of:
        0.0021279112 = product of:
          0.0042558224 = sum of:
            0.0042558224 = weight(_text_:a in 5604) [ClassicSimilarity], result of:
              0.0042558224 = score(doc=5604,freq=2.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.089176424 = fieldWeight in 5604, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5604)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  20. Sünkler, S.; Kerkmann, F.; Schultheiß, S.: Ok Google . the end of search as we know it : sprachgesteuerte Websuche im Test (2018) 0.00
    0.0010639556 = product of:
      0.0021279112 = sum of:
        0.0021279112 = product of:
          0.0042558224 = sum of:
            0.0042558224 = weight(_text_:a in 5626) [ClassicSimilarity], result of:
              0.0042558224 = score(doc=5626,freq=2.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.089176424 = fieldWeight in 5626, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5626)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a