Search (71 results, page 1 of 4)

  • × year_i:[2020 TO 2030}
  • × theme_ss:"Computerlinguistik"
  1. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.11
    0.106548965 = product of:
      0.45663843 = sum of:
        0.028466314 = product of:
          0.08539894 = sum of:
            0.08539894 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
              0.08539894 = score(doc=862,freq=2.0), product of:
                0.15195054 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.017922899 = queryNorm
                0.56201804 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=862)
          0.33333334 = coord(1/3)
        0.08539894 = weight(_text_:2f in 862) [ClassicSimilarity], result of:
          0.08539894 = score(doc=862,freq=2.0), product of:
            0.15195054 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.017922899 = queryNorm
            0.56201804 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
        0.08539894 = weight(_text_:2f in 862) [ClassicSimilarity], result of:
          0.08539894 = score(doc=862,freq=2.0), product of:
            0.15195054 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.017922899 = queryNorm
            0.56201804 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
        0.08539894 = weight(_text_:2f in 862) [ClassicSimilarity], result of:
          0.08539894 = score(doc=862,freq=2.0), product of:
            0.15195054 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.017922899 = queryNorm
            0.56201804 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
        0.08539894 = weight(_text_:2f in 862) [ClassicSimilarity], result of:
          0.08539894 = score(doc=862,freq=2.0), product of:
            0.15195054 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.017922899 = queryNorm
            0.56201804 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
        0.08539894 = weight(_text_:2f in 862) [ClassicSimilarity], result of:
          0.08539894 = score(doc=862,freq=2.0), product of:
            0.15195054 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.017922899 = queryNorm
            0.56201804 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
        0.0011773953 = product of:
          0.003532186 = sum of:
            0.003532186 = weight(_text_:a in 862) [ClassicSimilarity], result of:
              0.003532186 = score(doc=862,freq=10.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.1709182 = fieldWeight in 862, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=862)
          0.33333334 = coord(1/3)
      0.23333333 = coord(7/30)
    
    Abstract
    This research revisits the classic Turing test and compares recent large language models such as ChatGPT for their abilities to reproduce human-level comprehension and compelling text generation. Two task challenges- summary and question answering- prompt ChatGPT to produce original content (98-99%) from a single text entry and sequential questions initially posed by Turing in 1950. We score the original and generated content against the OpenAI GPT-2 Output Detector from 2019, and establish multiple cases where the generated content proves original and undetectable (98%). The question of a machine fooling a human judge recedes in this work relative to the question of "how would one prove it?" The original contribution of the work presents a metric and simple grammatical set for understanding the writing mechanics of chatbots in evaluating their readability and statistical clarity, engagement, delivery, overall quality, and plagiarism risks. While Turing's original prose scores at least 14% below the machine-generated output, whether an algorithm displays hints of Turing's true initial thoughts (the "Lovelace 2.0" test) remains unanswerable.
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
    Type
    a
  2. Barthel, J.; Ciesielski, R.: Regeln zu ChatGPT an Unis oft unklar : KI in der Bildung (2023) 0.00
    0.0034179986 = product of:
      0.051269975 = sum of:
        0.021220636 = product of:
          0.04244127 = sum of:
            0.04244127 = weight(_text_:29 in 925) [ClassicSimilarity], result of:
              0.04244127 = score(doc=925,freq=6.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.6731671 = fieldWeight in 925, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=925)
          0.5 = coord(1/2)
        0.030049339 = product of:
          0.04507401 = sum of:
            0.0026327355 = weight(_text_:a in 925) [ClassicSimilarity], result of:
              0.0026327355 = score(doc=925,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.12739488 = fieldWeight in 925, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=925)
            0.04244127 = weight(_text_:29 in 925) [ClassicSimilarity], result of:
              0.04244127 = score(doc=925,freq=6.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.6731671 = fieldWeight in 925, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=925)
          0.6666667 = coord(2/3)
      0.06666667 = coord(2/30)
    
    Date
    29. 3.2023 13:23:26
    29. 3.2023 13:29:19
    Type
    a
  3. Bager, J.: ¬Die Text-KI ChatGPT schreibt Fachtexte, Prosa, Gedichte und Programmcode (2023) 0.00
    0.0032965557 = product of:
      0.049448334 = sum of:
        0.035093244 = weight(_text_:medien in 835) [ClassicSimilarity], result of:
          0.035093244 = score(doc=835,freq=2.0), product of:
            0.084356464 = queryWeight, product of:
              4.7066307 = idf(docFreq=1085, maxDocs=44218)
              0.017922899 = queryNorm
            0.4160113 = fieldWeight in 835, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.7066307 = idf(docFreq=1085, maxDocs=44218)
              0.0625 = fieldNorm(doc=835)
        0.014355088 = product of:
          0.02153263 = sum of:
            0.0021061886 = weight(_text_:a in 835) [ClassicSimilarity], result of:
              0.0021061886 = score(doc=835,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.10191591 = fieldWeight in 835, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=835)
            0.019426443 = weight(_text_:22 in 835) [ClassicSimilarity], result of:
              0.019426443 = score(doc=835,freq=2.0), product of:
                0.06276294 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.017922899 = queryNorm
                0.30952093 = fieldWeight in 835, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=835)
          0.6666667 = coord(2/3)
      0.06666667 = coord(2/30)
    
    Abstract
    Das Forschungsunternehmen OpenAI hat ein neues Sprachmodell vorgestellt: ChatGPT. Es hat einen regelrechten Hype ausgelöst: Viele Nutzer haben Beispiele in den sozialen Medien gepostet, die die vielfältigen Fähigkeiten demonstrieren. Das darf aber nicht darüber hinwegtäuschen, dass auch ChatGPT einige gefährliche Schwächen hat, die für Sprachmodelle typisch sind.
    Date
    29.12.2022 18:22:55
    Type
    a
  4. Stieler, W.: Anzeichen von Bewusstsein bei ChatGPT und Co.? (2023) 0.00
    0.0032422296 = product of:
      0.04863344 = sum of:
        0.048019137 = weight(_text_:einzelne in 1047) [ClassicSimilarity], result of:
          0.048019137 = score(doc=1047,freq=2.0), product of:
            0.10548963 = queryWeight, product of:
              5.885746 = idf(docFreq=333, maxDocs=44218)
              0.017922899 = queryNorm
            0.45520246 = fieldWeight in 1047, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.885746 = idf(docFreq=333, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1047)
        6.14305E-4 = product of:
          0.001842915 = sum of:
            0.001842915 = weight(_text_:a in 1047) [ClassicSimilarity], result of:
              0.001842915 = score(doc=1047,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.089176424 = fieldWeight in 1047, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1047)
          0.33333334 = coord(1/3)
      0.06666667 = coord(2/30)
    
    Abstract
    Ein interdisziplinäres Forschungsteam hat eine Liste von Eigenschaften aufgestellt, die auf Bewusstsein deuten, und aktuelle KI-Systeme darauf abgeklopft. Ein interdisziplinäres Forscherteam hat ein Paper [https://arxiv.org/abs/2308.08708] veröffentlicht, das eine Liste von 14 "Indikatoren" für Bewusstsein enthält, die aus sechs aktuellen Theorien über das Bewusstsein stammen. Aktuelle KI-Modelle wie GPT-3, Palm-E oder AdA von Deepmind weisen demnach einzelne dieser Indikatoren auf. "Es spricht viel dafür, dass die meisten oder alle Bedingungen für das Bewusstsein, die von derzeitigen Theorien vorgeschlagenen Bedingungen für das Bewusstsein mit den bestehenden Techniken der KI erfüllt werden können", schreiben die Autoren. Zum Team gehörte auch der Deep-Learning-Pionier Yoshua Bengio von der Université de Montréal.
    Type
    a
  5. Leighton, T.: ChatGPT und Künstliche Intelligenz : Utopie oder Dystopie? (2023) 0.00
    0.0029829422 = product of:
      0.04474413 = sum of:
        0.043866552 = weight(_text_:medien in 908) [ClassicSimilarity], result of:
          0.043866552 = score(doc=908,freq=2.0), product of:
            0.084356464 = queryWeight, product of:
              4.7066307 = idf(docFreq=1085, maxDocs=44218)
              0.017922899 = queryNorm
            0.5200141 = fieldWeight in 908, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.7066307 = idf(docFreq=1085, maxDocs=44218)
              0.078125 = fieldNorm(doc=908)
        8.775785E-4 = product of:
          0.0026327355 = sum of:
            0.0026327355 = weight(_text_:a in 908) [ClassicSimilarity], result of:
              0.0026327355 = score(doc=908,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.12739488 = fieldWeight in 908, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=908)
          0.33333334 = coord(1/3)
      0.06666667 = coord(2/30)
    
    Series
    Telepolis / Kultur und Medien
    Type
    a
  6. Schneider, R.U.: Darf der Computer die Seminararbeit schreiben? (2023) 0.00
    0.0026998254 = product of:
      0.04049738 = sum of:
        0.039444286 = weight(_text_:neue in 1168) [ClassicSimilarity], result of:
          0.039444286 = score(doc=1168,freq=2.0), product of:
            0.07302189 = queryWeight, product of:
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.017922899 = queryNorm
            0.5401707 = fieldWeight in 1168, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.09375 = fieldNorm(doc=1168)
        0.0010530944 = product of:
          0.003159283 = sum of:
            0.003159283 = weight(_text_:a in 1168) [ClassicSimilarity], result of:
              0.003159283 = score(doc=1168,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.15287387 = fieldWeight in 1168, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=1168)
          0.33333334 = coord(1/3)
      0.06666667 = coord(2/30)
    
    Source
    Neue Zürcher Zeitung, 2023, vom 22.11.2023
    Type
    a
  7. Müller, P.: Text-Automat mit Tücken (2023) 0.00
    0.0024274043 = product of:
      0.036411062 = sum of:
        0.014702087 = product of:
          0.029404175 = sum of:
            0.029404175 = weight(_text_:29 in 481) [ClassicSimilarity], result of:
              0.029404175 = score(doc=481,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.46638384 = fieldWeight in 481, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=481)
          0.5 = coord(1/2)
        0.021708973 = product of:
          0.03256346 = sum of:
            0.003159283 = weight(_text_:a in 481) [ClassicSimilarity], result of:
              0.003159283 = score(doc=481,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.15287387 = fieldWeight in 481, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=481)
            0.029404175 = weight(_text_:29 in 481) [ClassicSimilarity], result of:
              0.029404175 = score(doc=481,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.46638384 = fieldWeight in 481, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=481)
          0.6666667 = coord(2/3)
      0.06666667 = coord(2/30)
    
    Source
    Pirmasenser Zeitung. Nr. 29 vom 03.02.2023, S.2
    Type
    a
  8. Lobo, S.: ¬Das Ende von Google, wie wir es kannten : Bessere Treffer durch ChatGPT (2022) 0.00
    0.0022740883 = product of:
      0.034111325 = sum of:
        0.03287024 = weight(_text_:neue in 852) [ClassicSimilarity], result of:
          0.03287024 = score(doc=852,freq=2.0), product of:
            0.07302189 = queryWeight, product of:
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.017922899 = queryNorm
            0.4501423 = fieldWeight in 852, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.078125 = fieldNorm(doc=852)
        0.0012410836 = product of:
          0.0037232507 = sum of:
            0.0037232507 = weight(_text_:a in 852) [ClassicSimilarity], result of:
              0.0037232507 = score(doc=852,freq=4.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.18016359 = fieldWeight in 852, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=852)
          0.33333334 = coord(1/3)
      0.06666667 = coord(2/30)
    
    Abstract
    Höchste Alarmstufe bei der weltgrößten Suchmaschine: Mit ChatGPT und künstlicher Intelligenz könnte eine neue Ära beginnen.
    Source
    https://www.spiegel.de/netzwelt/netzpolitik/bessere-treffer-durch-chatgpt-das-ende-von-google-wie-wir-es-kannten-kolumne-a-77820af6-51d7-4c03-b822-cf93094fd709
    Type
    a
  9. Hartnett, K.: Sind Sprachmodelle bald die besseren Mathematiker? (2023) 0.00
    0.0022498546 = product of:
      0.03374782 = sum of:
        0.03287024 = weight(_text_:neue in 988) [ClassicSimilarity], result of:
          0.03287024 = score(doc=988,freq=2.0), product of:
            0.07302189 = queryWeight, product of:
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.017922899 = queryNorm
            0.4501423 = fieldWeight in 988, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.078125 = fieldNorm(doc=988)
        8.775785E-4 = product of:
          0.0026327355 = sum of:
            0.0026327355 = weight(_text_:a in 988) [ClassicSimilarity], result of:
              0.0026327355 = score(doc=988,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.12739488 = fieldWeight in 988, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=988)
          0.33333334 = coord(1/3)
      0.06666667 = coord(2/30)
    
    Abstract
    Algorithmen wie ChatGPT tun sich selbst mit einfachen Rechenaufgaben noch immer schwer. Zwei neue Ansätze sollen das ändern - und könnten KI deutlich intelligenter machen.
    Type
    a
  10. Bischoff, M.: Was steckt hinter ChatGTP & Co? (2023) 0.00
    0.0016182694 = product of:
      0.02427404 = sum of:
        0.0098013915 = product of:
          0.019602783 = sum of:
            0.019602783 = weight(_text_:29 in 1013) [ClassicSimilarity], result of:
              0.019602783 = score(doc=1013,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.31092256 = fieldWeight in 1013, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1013)
          0.5 = coord(1/2)
        0.014472648 = product of:
          0.02170897 = sum of:
            0.0021061886 = weight(_text_:a in 1013) [ClassicSimilarity], result of:
              0.0021061886 = score(doc=1013,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.10191591 = fieldWeight in 1013, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1013)
            0.019602783 = weight(_text_:29 in 1013) [ClassicSimilarity], result of:
              0.019602783 = score(doc=1013,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.31092256 = fieldWeight in 1013, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1013)
          0.6666667 = coord(2/3)
      0.06666667 = coord(2/30)
    
    Date
    12. 4.2023 20:29:54
    Type
    a
  11. Dampz, N.: ChatGPT interpretiert jetzt auch Bilder : Neue Version (2023) 0.00
    0.0015495181 = product of:
      0.04648554 = sum of:
        0.04648554 = weight(_text_:neue in 874) [ClassicSimilarity], result of:
          0.04648554 = score(doc=874,freq=4.0), product of:
            0.07302189 = queryWeight, product of:
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.017922899 = queryNorm
            0.63659734 = fieldWeight in 874, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.074223 = idf(docFreq=2043, maxDocs=44218)
              0.078125 = fieldNorm(doc=874)
      0.033333335 = coord(1/30)
    
    Abstract
    Das kalifornische Unternehmen Open AI hat eine neue Version ihres Chatbots ChatGPT vorgestellt. Auffallendste Neuerung: Die Software, die mit Künstlicher Intelligenz funktioniert und bisher auf Text ausgerichtet war, interpretiert nun auch Bilder.
  12. Räwel, J.: Automatisierte Kommunikation (2023) 0.00
    0.0014914711 = product of:
      0.022372065 = sum of:
        0.021933276 = weight(_text_:medien in 909) [ClassicSimilarity], result of:
          0.021933276 = score(doc=909,freq=2.0), product of:
            0.084356464 = queryWeight, product of:
              4.7066307 = idf(docFreq=1085, maxDocs=44218)
              0.017922899 = queryNorm
            0.26000705 = fieldWeight in 909, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.7066307 = idf(docFreq=1085, maxDocs=44218)
              0.0390625 = fieldNorm(doc=909)
        4.3878925E-4 = product of:
          0.0013163678 = sum of:
            0.0013163678 = weight(_text_:a in 909) [ClassicSimilarity], result of:
              0.0013163678 = score(doc=909,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.06369744 = fieldWeight in 909, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=909)
          0.33333334 = coord(1/3)
      0.06666667 = coord(2/30)
    
    Series
    Telepolis / Kultur und Medien
    Type
    a
  13. Lund, B.D.; Wang, T.; Mannuru, N.R.; Nie, B.; Shimray, S.; Wang, Z.: ChatGPT and a new academic reality : artificial Intelligence-written research papers and the ethics of the large language models in scholarly publishing (2023) 0.00
    0.0013004817 = product of:
      0.019507226 = sum of:
        0.0073510436 = product of:
          0.014702087 = sum of:
            0.014702087 = weight(_text_:29 in 943) [ClassicSimilarity], result of:
              0.014702087 = score(doc=943,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.23319192 = fieldWeight in 943, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=943)
          0.5 = coord(1/2)
        0.012156183 = product of:
          0.018234273 = sum of:
            0.003532186 = weight(_text_:a in 943) [ClassicSimilarity], result of:
              0.003532186 = score(doc=943,freq=10.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.1709182 = fieldWeight in 943, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=943)
            0.014702087 = weight(_text_:29 in 943) [ClassicSimilarity], result of:
              0.014702087 = score(doc=943,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.23319192 = fieldWeight in 943, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=943)
          0.6666667 = coord(2/3)
      0.06666667 = coord(2/30)
    
    Abstract
    This article discusses OpenAI's ChatGPT, a generative pre-trained transformer, which uses natural language processing to fulfill text-based user requests (i.e., a "chatbot"). The history and principles behind ChatGPT and similar models are discussed. This technology is then discussed in relation to its potential impact on academia and scholarly research and publishing. ChatGPT is seen as a potential model for the automated preparation of essays and other types of scholarly manuscripts. Potential ethical issues that could arise with the emergence of large language models like GPT-3, the underlying technology behind ChatGPT, and its usage by academics and researchers, are discussed and situated within the context of broader advancements in artificial intelligence, machine learning, and natural language processing for research and scholarly publishing.
    Date
    19. 4.2023 19:29:44
    Type
    a
  14. Albrecht, I.: GPT-3: die Zukunft studentischer Hausarbeiten oder eine Bedrohung der wissenschaftlichen Integrität? (2023) 0.00
    0.0012137021 = product of:
      0.018205531 = sum of:
        0.0073510436 = product of:
          0.014702087 = sum of:
            0.014702087 = weight(_text_:29 in 881) [ClassicSimilarity], result of:
              0.014702087 = score(doc=881,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.23319192 = fieldWeight in 881, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=881)
          0.5 = coord(1/2)
        0.010854486 = product of:
          0.01628173 = sum of:
            0.0015796415 = weight(_text_:a in 881) [ClassicSimilarity], result of:
              0.0015796415 = score(doc=881,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.07643694 = fieldWeight in 881, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=881)
            0.014702087 = weight(_text_:29 in 881) [ClassicSimilarity], result of:
              0.014702087 = score(doc=881,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.23319192 = fieldWeight in 881, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=881)
          0.6666667 = coord(2/3)
      0.06666667 = coord(2/30)
    
    Date
    28. 1.2022 11:05:29
    Type
    a
  15. Pepper, S.; Arnaud, P.J.L.: Absolutely PHAB : toward a general model of associative relations (2020) 0.00
    0.001137923 = product of:
      0.017068844 = sum of:
        0.00612587 = product of:
          0.01225174 = sum of:
            0.01225174 = weight(_text_:29 in 103) [ClassicSimilarity], result of:
              0.01225174 = score(doc=103,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.19432661 = fieldWeight in 103, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=103)
          0.5 = coord(1/2)
        0.010942973 = product of:
          0.01641446 = sum of:
            0.0041627204 = weight(_text_:a in 103) [ClassicSimilarity], result of:
              0.0041627204 = score(doc=103,freq=20.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.20142901 = fieldWeight in 103, product of:
                  4.472136 = tf(freq=20.0), with freq of:
                    20.0 = termFreq=20.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=103)
            0.01225174 = weight(_text_:29 in 103) [ClassicSimilarity], result of:
              0.01225174 = score(doc=103,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.19432661 = fieldWeight in 103, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=103)
          0.6666667 = coord(2/3)
      0.06666667 = coord(2/30)
    
    Abstract
    There have been many attempts at classifying the semantic modification relations (R) of N + N compounds but this work has not led to the acceptance of a definitive scheme, so that devising a reusable classification is a worthwhile aim. The scope of this undertaking is extended to other binominal lexemes, i.e. units that contain two thing-morphemes without explicitly stating R, like prepositional units, N + relational adjective units, etc. The 25-relation taxonomy of Bourque (2014) was tested against over 15,000 binominal lexemes from 106 languages and extended to a 29-relation scheme ("Bourque2") through the introduction of two new reversible relations. Bourque2 is then mapped onto Hatcher's (1960) four-relation scheme (extended by the addition of a fifth relation, similarity , as "Hatcher2"). This results in a two-tier system usable at different degrees of granularities. On account of its semantic proximity to compounding, metonymy is then taken into account, following Janda's (2011) suggestion that it plays a role in word formation; Peirsman and Geeraerts' (2006) inventory of 23 metonymic patterns is mapped onto Bourque2, confirming the identity of metonymic and binominal modification relations. Finally, Blank's (2003) and Koch's (2001) work on lexical semantics justifies the addition to the scheme of a third, superordinate level which comprises the three Aristotelean principles of similarity, contiguity and contrast.
    Type
    a
  16. Laparra, E.; Binford-Walsh, A.; Emerson, K.; Miller, M.L.; López-Hoffman, L.; Currim, F.; Bethard, S.: Addressing structural hurdles for metadata extraction from environmental impact statements (2023) 0.00
    0.0010699236 = product of:
      0.016048854 = sum of:
        0.00612587 = product of:
          0.01225174 = sum of:
            0.01225174 = weight(_text_:29 in 1042) [ClassicSimilarity], result of:
              0.01225174 = score(doc=1042,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.19432661 = fieldWeight in 1042, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1042)
          0.5 = coord(1/2)
        0.009922984 = product of:
          0.014884476 = sum of:
            0.0026327355 = weight(_text_:a in 1042) [ClassicSimilarity], result of:
              0.0026327355 = score(doc=1042,freq=8.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.12739488 = fieldWeight in 1042, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1042)
            0.01225174 = weight(_text_:29 in 1042) [ClassicSimilarity], result of:
              0.01225174 = score(doc=1042,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.19432661 = fieldWeight in 1042, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1042)
          0.6666667 = coord(2/3)
      0.06666667 = coord(2/30)
    
    Abstract
    Natural language processing techniques can be used to analyze the linguistic content of a document to extract missing pieces of metadata. However, accurate metadata extraction may not depend solely on the linguistics, but also on structural problems such as extremely large documents, unordered multi-file documents, and inconsistency in manually labeled metadata. In this work, we start from two standard machine learning solutions to extract pieces of metadata from Environmental Impact Statements, environmental policy documents that are regularly produced under the US National Environmental Policy Act of 1969. We present a series of experiments where we evaluate how these standard approaches are affected by different issues derived from real-world data. We find that metadata extraction can be strongly influenced by nonlinguistic factors such as document length and volume ordering and that the standard machine learning solutions often do not scale well to long documents. We demonstrate how such solutions can be better adapted to these scenarios, and conclude with suggestions for other NLP practitioners cataloging large document collections.
    Date
    29. 8.2023 19:21:01
    Type
    a
  17. Hahn, U.: Automatische Sprachverarbeitung (2023) 0.00
    0.0010317749 = product of:
      0.015476623 = sum of:
        0.014862318 = weight(_text_:u in 790) [ClassicSimilarity], result of:
          0.014862318 = score(doc=790,freq=2.0), product of:
            0.058687534 = queryWeight, product of:
              3.2744443 = idf(docFreq=4547, maxDocs=44218)
              0.017922899 = queryNorm
            0.25324488 = fieldWeight in 790, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2744443 = idf(docFreq=4547, maxDocs=44218)
              0.0546875 = fieldNorm(doc=790)
        6.14305E-4 = product of:
          0.001842915 = sum of:
            0.001842915 = weight(_text_:a in 790) [ClassicSimilarity], result of:
              0.001842915 = score(doc=790,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.089176424 = fieldWeight in 790, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=790)
          0.33333334 = coord(1/3)
      0.06666667 = coord(2/30)
    
    Type
    a
  18. ¬Der Student aus dem Computer (2023) 0.00
    8.373802E-4 = product of:
      0.025121404 = sum of:
        0.025121404 = product of:
          0.037682105 = sum of:
            0.00368583 = weight(_text_:a in 1079) [ClassicSimilarity], result of:
              0.00368583 = score(doc=1079,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.17835285 = fieldWeight in 1079, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1079)
            0.033996277 = weight(_text_:22 in 1079) [ClassicSimilarity], result of:
              0.033996277 = score(doc=1079,freq=2.0), product of:
                0.06276294 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.017922899 = queryNorm
                0.5416616 = fieldWeight in 1079, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1079)
          0.6666667 = coord(2/3)
      0.033333335 = coord(1/30)
    
    Date
    27. 1.2023 16:22:55
    Type
    a
  19. Thomas, I.S.; Wang, J.; GPT-3: Was euch zu Menschen macht : Antworten einer künstlichen Intelligenz auf die großen Fragen des Lebens (2022) 0.00
    8.1678265E-4 = product of:
      0.012251739 = sum of:
        0.0073510436 = product of:
          0.014702087 = sum of:
            0.014702087 = weight(_text_:29 in 878) [ClassicSimilarity], result of:
              0.014702087 = score(doc=878,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.23319192 = fieldWeight in 878, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=878)
          0.5 = coord(1/2)
        0.0049006958 = product of:
          0.014702087 = sum of:
            0.014702087 = weight(_text_:29 in 878) [ClassicSimilarity], result of:
              0.014702087 = score(doc=878,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.23319192 = fieldWeight in 878, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=878)
          0.33333334 = coord(1/3)
      0.06666667 = coord(2/30)
    
    Date
    7. 1.2023 18:41:29
  20. Donath, A.: Nutzungsverbote für ChatGPT (2023) 0.00
    8.091347E-4 = product of:
      0.01213702 = sum of:
        0.0049006958 = product of:
          0.0098013915 = sum of:
            0.0098013915 = weight(_text_:29 in 877) [ClassicSimilarity], result of:
              0.0098013915 = score(doc=877,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.15546128 = fieldWeight in 877, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03125 = fieldNorm(doc=877)
          0.5 = coord(1/2)
        0.007236324 = product of:
          0.010854485 = sum of:
            0.0010530943 = weight(_text_:a in 877) [ClassicSimilarity], result of:
              0.0010530943 = score(doc=877,freq=2.0), product of:
                0.020665944 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.017922899 = queryNorm
                0.050957955 = fieldWeight in 877, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.03125 = fieldNorm(doc=877)
            0.0098013915 = weight(_text_:29 in 877) [ClassicSimilarity], result of:
              0.0098013915 = score(doc=877,freq=2.0), product of:
                0.063047156 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.017922899 = queryNorm
                0.15546128 = fieldWeight in 877, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03125 = fieldNorm(doc=877)
          0.6666667 = coord(2/3)
      0.06666667 = coord(2/30)
    
    Content
    Milliardenbewertung für ChatGPT OpenAI, das Chatbot ChatGPT betreibt, befindet sich laut einem Bericht des Wall Street Journals in Gesprächen zu einem Aktienverkauf. Das WSJ meldete, der mögliche Verkauf der Aktien würde die Bewertung von OpenAI auf 29 Milliarden US-Dollar anheben. Sorgen auch in Brandenburg Der brandenburgische SPD-Abgeordnete Erik Stohn stellte mit Hilfe von ChatGPT eine Kleine Anfrage an den Brandenburger Landtag, in der er fragte, wie die Landesregierung sicherstelle, dass Studierende bei maschinell erstellten Texten gerecht beurteilt und benotet würden. Er fragte auch nach Maßnahmen, die ergriffen worden seien, um sicherzustellen, dass maschinell erstellte Texte nicht in betrügerischer Weise von Studierenden bei der Bewertung von Studienleistungen verwendet werden könnten.

Languages

  • e 38
  • d 32

Types

  • a 57
  • el 36
  • p 7
  • m 1
  • x 1
  • More… Less…