Search (1 results, page 1 of 1)

  • × author_ss:"Child, R."
  • × type_ss:"el"
  1. Radford, A.; Wu, J.; Child, R.; Luan, D.; Amode, D.; Sutskever, I.: Language models are unsupervised multitask learners 0.03
    0.031724695 = product of:
      0.06344939 = sum of:
        0.06344939 = product of:
          0.12689878 = sum of:
            0.12689878 = weight(_text_:plus in 871) [ClassicSimilarity], result of:
              0.12689878 = score(doc=871,freq=2.0), product of:
                0.3101809 = queryWeight, product of:
                  6.1714344 = idf(docFreq=250, maxDocs=44218)
                  0.05026075 = queryNorm
                0.40911216 = fieldWeight in 871, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  6.1714344 = idf(docFreq=250, maxDocs=44218)
                  0.046875 = fieldNorm(doc=871)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Natural language processing tasks, such as question answering, machine translation, reading comprehension, and summarization, are typically approached with supervised learning on task-specific datasets. We demonstrate that language models begin to learn these tasks without any explicit supervision when trained on a new dataset of millions of webpages called WebText. When conditioned on a document plus questions, the answers generated by the language model reach 55 F1 on the CoQA dataset - matching or exceeding the performance of 3 out of 4 baseline systems without using the 127,000+ training examples. The capacity of the language model is essential to the success of zero-shot task transfer and increasing it improves performance in a log-linear fashion across tasks. Our largest model, GPT-2, is a 1.5B parameter Transformer that achieves state of the art results on 7 out of 8 tested language modeling datasets in a zero-shot setting but still underfits WebText. Samples from the model reflect these improvements and contain coherent paragraphs of text. These findings suggest a promising path towards building language processing systems which learn to perform tasks from their naturally occurring demonstrations.