Search (17 results, page 1 of 1)

  • × theme_ss:"Computerlinguistik"
  • × type_ss:"a"
  • × year_i:[2020 TO 2030}
  1. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.02
    0.017030947 = product of:
      0.06812379 = sum of:
        0.06812379 = product of:
          0.20437136 = sum of:
            0.20437136 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
              0.20437136 = score(doc=862,freq=2.0), product of:
                0.36363843 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.04289195 = queryNorm
                0.56201804 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=862)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
  2. Barthel, J.; Ciesielski, R.: Regeln zu ChatGPT an Unis oft unklar : KI in der Bildung (2023) 0.01
    0.008463981 = product of:
      0.033855923 = sum of:
        0.033855923 = product of:
          0.10156777 = sum of:
            0.10156777 = weight(_text_:29 in 925) [ClassicSimilarity], result of:
              0.10156777 = score(doc=925,freq=6.0), product of:
                0.15088047 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04289195 = queryNorm
                0.6731671 = fieldWeight in 925, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=925)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    29. 3.2023 13:23:26
    29. 3.2023 13:29:19
  3. ¬Der Student aus dem Computer (2023) 0.01
    0.0067798113 = product of:
      0.027119245 = sum of:
        0.027119245 = product of:
          0.08135773 = sum of:
            0.08135773 = weight(_text_:22 in 1079) [ClassicSimilarity], result of:
              0.08135773 = score(doc=1079,freq=2.0), product of:
                0.15020029 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04289195 = queryNorm
                0.5416616 = fieldWeight in 1079, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1079)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    27. 1.2023 16:22:55
  4. Müller, P.: Text-Automat mit Tücken (2023) 0.01
    0.005864018 = product of:
      0.023456072 = sum of:
        0.023456072 = product of:
          0.070368215 = sum of:
            0.070368215 = weight(_text_:29 in 481) [ClassicSimilarity], result of:
              0.070368215 = score(doc=481,freq=2.0), product of:
                0.15088047 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04289195 = queryNorm
                0.46638384 = fieldWeight in 481, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=481)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Source
    Pirmasenser Zeitung. Nr. 29 vom 03.02.2023, S.2
  5. Zaitseva, E.M.: Developing linguistic tools of thematic search in library information systems (2023) 0.00
    0.004169932 = product of:
      0.016679728 = sum of:
        0.016679728 = product of:
          0.050039183 = sum of:
            0.050039183 = weight(_text_:systems in 1187) [ClassicSimilarity], result of:
              0.050039183 = score(doc=1187,freq=10.0), product of:
                0.13181444 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.04289195 = queryNorm
                0.37961838 = fieldWeight in 1187, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1187)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    Within the R&D program "Information support of research by scientists and specialists on the basis of RNPLS&T Open Archive - the system of scientific knowledge aggregation", the RNPLS&T analyzes the use of linguistic tools of thematic search in the modern library information systems and the prospects for their development. The author defines the key common characteristics of e-catalogs of the largest Russian libraries revealed at the first stage of the analysis. Based on the specified common characteristics and detailed comparison analysis, the author outlines and substantiates the vectors for enhancing search inter faces of e-catalogs. The focus is made on linguistic tools of thematic search in library information systems; the key vectors are suggested: use of thematic search at different search levels with the clear-cut level differentiation; use of combined functionality within thematic search system; implementation of classification search in all e-catalogs; hierarchical representation of classifications; use of the matching systems for classification information retrieval languages, and in the long term classification and verbal information retrieval languages, and various verbal information retrieval languages. The author formulates practical recommendations to improve thematic search in library information systems.
  6. Bischoff, M.: Was steckt hinter ChatGTP & Co? (2023) 0.00
    0.0039093453 = product of:
      0.015637381 = sum of:
        0.015637381 = product of:
          0.04691214 = sum of:
            0.04691214 = weight(_text_:29 in 1013) [ClassicSimilarity], result of:
              0.04691214 = score(doc=1013,freq=2.0), product of:
                0.15088047 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04289195 = queryNorm
                0.31092256 = fieldWeight in 1013, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1013)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    12. 4.2023 20:29:54
  7. Morris, V.: Automated language identification of bibliographic resources (2020) 0.00
    0.0038741778 = product of:
      0.015496711 = sum of:
        0.015496711 = product of:
          0.046490133 = sum of:
            0.046490133 = weight(_text_:22 in 5749) [ClassicSimilarity], result of:
              0.046490133 = score(doc=5749,freq=2.0), product of:
                0.15020029 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04289195 = queryNorm
                0.30952093 = fieldWeight in 5749, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5749)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    2. 3.2020 19:04:22
  8. Bager, J.: ¬Die Text-KI ChatGPT schreibt Fachtexte, Prosa, Gedichte und Programmcode (2023) 0.00
    0.0038741778 = product of:
      0.015496711 = sum of:
        0.015496711 = product of:
          0.046490133 = sum of:
            0.046490133 = weight(_text_:22 in 835) [ClassicSimilarity], result of:
              0.046490133 = score(doc=835,freq=2.0), product of:
                0.15020029 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04289195 = queryNorm
                0.30952093 = fieldWeight in 835, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=835)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    29.12.2022 18:22:55
  9. Rieger, F.: Lügende Computer (2023) 0.00
    0.0038741778 = product of:
      0.015496711 = sum of:
        0.015496711 = product of:
          0.046490133 = sum of:
            0.046490133 = weight(_text_:22 in 912) [ClassicSimilarity], result of:
              0.046490133 = score(doc=912,freq=2.0), product of:
                0.15020029 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04289195 = queryNorm
                0.30952093 = fieldWeight in 912, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=912)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    16. 3.2023 19:22:55
  10. Albrecht, I.: GPT-3: die Zukunft studentischer Hausarbeiten oder eine Bedrohung der wissenschaftlichen Integrität? (2023) 0.00
    0.002932009 = product of:
      0.011728036 = sum of:
        0.011728036 = product of:
          0.035184108 = sum of:
            0.035184108 = weight(_text_:29 in 881) [ClassicSimilarity], result of:
              0.035184108 = score(doc=881,freq=2.0), product of:
                0.15088047 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04289195 = queryNorm
                0.23319192 = fieldWeight in 881, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=881)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    28. 1.2022 11:05:29
  11. Lund, B.D.; Wang, T.; Mannuru, N.R.; Nie, B.; Shimray, S.; Wang, Z.: ChatGPT and a new academic reality : artificial Intelligence-written research papers and the ethics of the large language models in scholarly publishing (2023) 0.00
    0.002932009 = product of:
      0.011728036 = sum of:
        0.011728036 = product of:
          0.035184108 = sum of:
            0.035184108 = weight(_text_:29 in 943) [ClassicSimilarity], result of:
              0.035184108 = score(doc=943,freq=2.0), product of:
                0.15088047 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04289195 = queryNorm
                0.23319192 = fieldWeight in 943, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=943)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    19. 4.2023 19:29:44
  12. Escolano, C.; Costa-Jussà, M.R.; Fonollosa, J.A.: From bilingual to multilingual neural-based machine translation by incremental training (2021) 0.00
    0.0026372964 = product of:
      0.010549186 = sum of:
        0.010549186 = product of:
          0.031647556 = sum of:
            0.031647556 = weight(_text_:systems in 97) [ClassicSimilarity], result of:
              0.031647556 = score(doc=97,freq=4.0), product of:
                0.13181444 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.04289195 = queryNorm
                0.24009174 = fieldWeight in 97, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=97)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    A common intermediate language representation in neural machine translation can be used to extend bilingual systems by incremental training. We propose a new architecture based on introducing an interlingual loss as an additional training objective. By adding and forcing this interlingual loss, we can train multiple encoders and decoders for each language, sharing among them a common intermediate representation. Translation results on the low-resource tasks (Turkish-English and Kazakh-English tasks) show a BLEU improvement of up to 2.8 points. However, results on a larger dataset (Russian-English and Kazakh-English) show BLEU losses of a similar amount. While our system provides improvements only for the low-resource tasks in terms of translation quality, our system is capable of quickly deploying new language pairs without the need to retrain the rest of the system, which may be a game changer in some situations. Specifically, what is most relevant regarding our architecture is that it is capable of: reducing the number of production systems, with respect to the number of languages, from quadratic to linear; incrementally adding a new language to the system without retraining the languages already there; and allowing for translations from the new language to all the others present in the system.
  13. Aizawa, A.; Kohlhase, M.: Mathematical information retrieval (2021) 0.00
    0.0026107903 = product of:
      0.010443161 = sum of:
        0.010443161 = product of:
          0.031329483 = sum of:
            0.031329483 = weight(_text_:systems in 667) [ClassicSimilarity], result of:
              0.031329483 = score(doc=667,freq=2.0), product of:
                0.13181444 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.04289195 = queryNorm
                0.23767869 = fieldWeight in 667, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=667)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    We present an overview of the NTCIR Math Tasks organized during NTCIR-10, 11, and 12. These tasks are primarily dedicated to techniques for searching mathematical content with formula expressions. In this chapter, we first summarize the task design and introduce test collections generated in the tasks. We also describe the features and main challenges of mathematical information retrieval systems and discuss future perspectives in the field.
  14. Collard, J.; Paiva, V. de; Fong, B.; Subrahmanian, E.: Extracting mathematical concepts from text (2022) 0.00
    0.0026107903 = product of:
      0.010443161 = sum of:
        0.010443161 = product of:
          0.031329483 = sum of:
            0.031329483 = weight(_text_:systems in 668) [ClassicSimilarity], result of:
              0.031329483 = score(doc=668,freq=2.0), product of:
                0.13181444 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.04289195 = queryNorm
                0.23767869 = fieldWeight in 668, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=668)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    We investigate different systems for extracting mathematical entities from English texts in the mathematical field of category theory as a first step for constructing a mathematical knowledge graph. We consider four different term extractors and compare their results. This small experiment showcases some of the issues with the construction and evaluation of terms extracted from noisy domain text. We also make available two open corpora in research mathematics, in particular in category theory: a small corpus of 755 abstracts from the journal TAC (3188 sentences), and a larger corpus from the nLab community wiki (15,000 sentences).
  15. Pepper, S.; Arnaud, P.J.L.: Absolutely PHAB : toward a general model of associative relations (2020) 0.00
    0.002443341 = product of:
      0.009773364 = sum of:
        0.009773364 = product of:
          0.029320091 = sum of:
            0.029320091 = weight(_text_:29 in 103) [ClassicSimilarity], result of:
              0.029320091 = score(doc=103,freq=2.0), product of:
                0.15088047 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04289195 = queryNorm
                0.19432661 = fieldWeight in 103, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=103)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    There have been many attempts at classifying the semantic modification relations (R) of N + N compounds but this work has not led to the acceptance of a definitive scheme, so that devising a reusable classification is a worthwhile aim. The scope of this undertaking is extended to other binominal lexemes, i.e. units that contain two thing-morphemes without explicitly stating R, like prepositional units, N + relational adjective units, etc. The 25-relation taxonomy of Bourque (2014) was tested against over 15,000 binominal lexemes from 106 languages and extended to a 29-relation scheme ("Bourque2") through the introduction of two new reversible relations. Bourque2 is then mapped onto Hatcher's (1960) four-relation scheme (extended by the addition of a fifth relation, similarity , as "Hatcher2"). This results in a two-tier system usable at different degrees of granularities. On account of its semantic proximity to compounding, metonymy is then taken into account, following Janda's (2011) suggestion that it plays a role in word formation; Peirsman and Geeraerts' (2006) inventory of 23 metonymic patterns is mapped onto Bourque2, confirming the identity of metonymic and binominal modification relations. Finally, Blank's (2003) and Koch's (2001) work on lexical semantics justifies the addition to the scheme of a third, superordinate level which comprises the three Aristotelean principles of similarity, contiguity and contrast.
  16. Laparra, E.; Binford-Walsh, A.; Emerson, K.; Miller, M.L.; López-Hoffman, L.; Currim, F.; Bethard, S.: Addressing structural hurdles for metadata extraction from environmental impact statements (2023) 0.00
    0.002443341 = product of:
      0.009773364 = sum of:
        0.009773364 = product of:
          0.029320091 = sum of:
            0.029320091 = weight(_text_:29 in 1042) [ClassicSimilarity], result of:
              0.029320091 = score(doc=1042,freq=2.0), product of:
                0.15088047 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04289195 = queryNorm
                0.19432661 = fieldWeight in 1042, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1042)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    29. 8.2023 19:21:01
  17. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; Agarwal, S.; Herbert-Voss, A.; Krueger, G.; Henighan, T.; Child, R.; Ramesh, A.; Ziegler, D.M.; Wu, J.; Winter, C.; Hesse, C.; Chen, M.; Sigler, E.; Litwin, M.; Gray, S.; Chess, B.; Clark, J.; Berner, C.; McCandlish, S.; Radford, A.; Sutskever, I.; Amodei, D.: Language models are few-shot learners (2020) 0.00
    0.00149188 = product of:
      0.00596752 = sum of:
        0.00596752 = product of:
          0.01790256 = sum of:
            0.01790256 = weight(_text_:systems in 872) [ClassicSimilarity], result of:
              0.01790256 = score(doc=872,freq=2.0), product of:
                0.13181444 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.04289195 = queryNorm
                0.1358164 = fieldWeight in 872, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.03125 = fieldNorm(doc=872)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    Recent work has demonstrated substantial gains on many NLP tasks and benchmarks by pre-training on a large corpus of text followed by fine-tuning on a specific task. While typically task-agnostic in architecture, this method still requires task-specific fine-tuning datasets of thousands or tens of thousands of examples. By contrast, humans can generally perform a new language task from only a few examples or from simple instructions - something which current NLP systems still largely struggle to do. Here we show that scaling up language models greatly improves task-agnostic, few-shot performance, sometimes even reaching competitiveness with prior state-of-the-art fine-tuning approaches. Specifically, we train GPT-3, an autoregressive language model with 175 billion parameters, 10x more than any previous non-sparse language model, and test its performance in the few-shot setting. For all tasks, GPT-3 is applied without any gradient updates or fine-tuning, with tasks and few-shot demonstrations specified purely via text interaction with the model. GPT-3 achieves strong performance on many NLP datasets, including translation, question-answering, and cloze tasks, as well as several tasks that require on-the-fly reasoning or domain adaptation, such as unscrambling words, using a novel word in a sentence, or performing 3-digit arithmetic. At the same time, we also identify some datasets where GPT-3's few-shot learning still struggles, as well as some datasets where GPT-3 faces methodological issues related to training on large web corpora. Finally, we find that GPT-3 can generate samples of news articles which human evaluators have difficulty distinguishing from articles written by humans. We discuss broader societal impacts of this finding and of GPT-3 in general.