Search (1 results, page 1 of 1)

  • × author_ss:"Child, R."
  • × theme_ss:"Computerlinguistik"
  • × type_ss:"el"
  1. Radford, A.; Wu, J.; Child, R.; Luan, D.; Amode, D.; Sutskever, I.: Language models are unsupervised multitask learners 0.01
    0.0075771073 = product of:
      0.037885536 = sum of:
        0.037885536 = weight(_text_:7 in 871) [ClassicSimilarity], result of:
          0.037885536 = score(doc=871,freq=2.0), product of:
            0.17251469 = queryWeight, product of:
              3.3127685 = idf(docFreq=4376, maxDocs=44218)
              0.052075688 = queryNorm
            0.21960759 = fieldWeight in 871, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3127685 = idf(docFreq=4376, maxDocs=44218)
              0.046875 = fieldNorm(doc=871)
      0.2 = coord(1/5)
    
    Abstract
    Natural language processing tasks, such as question answering, machine translation, reading comprehension, and summarization, are typically approached with supervised learning on task-specific datasets. We demonstrate that language models begin to learn these tasks without any explicit supervision when trained on a new dataset of millions of webpages called WebText. When conditioned on a document plus questions, the answers generated by the language model reach 55 F1 on the CoQA dataset - matching or exceeding the performance of 3 out of 4 baseline systems without using the 127,000+ training examples. The capacity of the language model is essential to the success of zero-shot task transfer and increasing it improves performance in a log-linear fashion across tasks. Our largest model, GPT-2, is a 1.5B parameter Transformer that achieves state of the art results on 7 out of 8 tested language modeling datasets in a zero-shot setting but still underfits WebText. Samples from the model reflect these improvements and contain coherent paragraphs of text. These findings suggest a promising path towards building language processing systems which learn to perform tasks from their naturally occurring demonstrations.