Search (68 results, page 1 of 4)

  • × theme_ss:"Computerlinguistik"
  1. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.32
    0.31868097 = product of:
      0.47802144 = sum of:
        0.065880254 = product of:
          0.19764076 = sum of:
            0.19764076 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
              0.19764076 = score(doc=562,freq=2.0), product of:
                0.35166267 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.041479383 = queryNorm
                0.56201804 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
        0.19764076 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.19764076 = score(doc=562,freq=2.0), product of:
            0.35166267 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041479383 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.19764076 = weight(_text_:2f in 562) [ClassicSimilarity], result of:
          0.19764076 = score(doc=562,freq=2.0), product of:
            0.35166267 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041479383 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.01685965 = product of:
          0.0337193 = sum of:
            0.0337193 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
              0.0337193 = score(doc=562,freq=2.0), product of:
                0.14525373 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.041479383 = queryNorm
                0.23214069 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.5 = coord(1/2)
      0.6666667 = coord(4/6)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  2. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.23
    0.2305809 = product of:
      0.4611618 = sum of:
        0.065880254 = product of:
          0.19764076 = sum of:
            0.19764076 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
              0.19764076 = score(doc=862,freq=2.0), product of:
                0.35166267 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.041479383 = queryNorm
                0.56201804 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=862)
          0.33333334 = coord(1/3)
        0.19764076 = weight(_text_:2f in 862) [ClassicSimilarity], result of:
          0.19764076 = score(doc=862,freq=2.0), product of:
            0.35166267 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041479383 = queryNorm
            0.56201804 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
        0.19764076 = weight(_text_:2f in 862) [ClassicSimilarity], result of:
          0.19764076 = score(doc=862,freq=2.0), product of:
            0.35166267 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041479383 = queryNorm
            0.56201804 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
      0.5 = coord(3/6)
    
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
  3. Huo, W.: Automatic multi-word term extraction and its application to Web-page summarization (2012) 0.21
    0.20607059 = product of:
      0.41214117 = sum of:
        0.19764076 = weight(_text_:2f in 563) [ClassicSimilarity], result of:
          0.19764076 = score(doc=563,freq=2.0), product of:
            0.35166267 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041479383 = queryNorm
            0.56201804 = fieldWeight in 563, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=563)
        0.19764076 = weight(_text_:2f in 563) [ClassicSimilarity], result of:
          0.19764076 = score(doc=563,freq=2.0), product of:
            0.35166267 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.041479383 = queryNorm
            0.56201804 = fieldWeight in 563, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=563)
        0.01685965 = product of:
          0.0337193 = sum of:
            0.0337193 = weight(_text_:22 in 563) [ClassicSimilarity], result of:
              0.0337193 = score(doc=563,freq=2.0), product of:
                0.14525373 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.041479383 = queryNorm
                0.23214069 = fieldWeight in 563, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=563)
          0.5 = coord(1/2)
      0.5 = coord(3/6)
    
    Content
    A Thesis presented to The University of Guelph In partial fulfilment of requirements for the degree of Master of Science in Computer Science. Vgl. Unter: http://www.inf.ufrgs.br%2F~ceramisch%2Fdownload_files%2Fpublications%2F2009%2Fp01.pdf.
    Date
    10. 1.2013 19:22:47
  4. Liddy, E.D.: Natural language processing for information retrieval and knowledge discovery (1998) 0.08
    0.07518271 = product of:
      0.22554813 = sum of:
        0.1057348 = weight(_text_:21st in 2345) [ClassicSimilarity], result of:
          0.1057348 = score(doc=2345,freq=2.0), product of:
            0.2381352 = queryWeight, product of:
              5.74105 = idf(docFreq=385, maxDocs=44218)
              0.041479383 = queryNorm
            0.44401163 = fieldWeight in 2345, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.74105 = idf(docFreq=385, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2345)
        0.11981333 = sum of:
          0.080474146 = weight(_text_:century in 2345) [ClassicSimilarity], result of:
            0.080474146 = score(doc=2345,freq=2.0), product of:
              0.20775084 = queryWeight, product of:
                5.0085325 = idf(docFreq=802, maxDocs=44218)
                0.041479383 = queryNorm
              0.38735893 = fieldWeight in 2345, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.0085325 = idf(docFreq=802, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2345)
          0.039339185 = weight(_text_:22 in 2345) [ClassicSimilarity], result of:
            0.039339185 = score(doc=2345,freq=2.0), product of:
              0.14525373 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041479383 = queryNorm
              0.2708308 = fieldWeight in 2345, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2345)
      0.33333334 = coord(2/6)
    
    Date
    22. 9.1997 19:16:05
    Source
    Visualizing subject access for 21st century information resources: Papers presented at the 1997 Clinic on Library Applications of Data Processing, 2-4 Mar 1997, Graduate School of Library and Information Science, University of Illinois at Urbana-Champaign. Ed.: P.A. Cochrane et al
  5. Mustafa el Hadi, W.: Terminology & information retrieval : new tools for new needs. Integration of knowledge across boundaries (2003) 0.05
    0.04646818 = product of:
      0.13940454 = sum of:
        0.09062983 = weight(_text_:21st in 2688) [ClassicSimilarity], result of:
          0.09062983 = score(doc=2688,freq=2.0), product of:
            0.2381352 = queryWeight, product of:
              5.74105 = idf(docFreq=385, maxDocs=44218)
              0.041479383 = queryNorm
            0.3805814 = fieldWeight in 2688, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.74105 = idf(docFreq=385, maxDocs=44218)
              0.046875 = fieldNorm(doc=2688)
        0.048774697 = product of:
          0.097549394 = sum of:
            0.097549394 = weight(_text_:century in 2688) [ClassicSimilarity], result of:
              0.097549394 = score(doc=2688,freq=4.0), product of:
                0.20775084 = queryWeight, product of:
                  5.0085325 = idf(docFreq=802, maxDocs=44218)
                  0.041479383 = queryNorm
                0.46954992 = fieldWeight in 2688, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.0085325 = idf(docFreq=802, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2688)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    The radical changes in information and communication techniques at the end of the 20th century have significantly modified the function of terminology and its applications in all forms of communication. The introduction of new mediums has deeply changed the possibilities of distribution of scientific information. What in this situation is the role of terminology and its practical applications? What is the place for multiple functions of terminology in the communication society? What is the impact of natural language (NLP) techniques used in its processing and management? In this article we will focus an the possibilities NLP techniques offer and how they can be directed towards the satisfaction of the newly expressed needs.
    Source
    Challenges in knowledge representation and organization for the 21st century: Integration of knowledge across boundaries. Proceedings of the 7th ISKO International Conference Granada, Spain, July 10-13, 2002. Ed.: M. López-Huertas
  6. Hutchins, J.: From first conception to first demonstration : the nascent years of machine translation, 1947-1954. A chronology (1997) 0.04
    0.04242566 = product of:
      0.12727697 = sum of:
        0.099177554 = weight(_text_:history in 1463) [ClassicSimilarity], result of:
          0.099177554 = score(doc=1463,freq=2.0), product of:
            0.19296135 = queryWeight, product of:
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.041479383 = queryNorm
            0.5139763 = fieldWeight in 1463, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.078125 = fieldNorm(doc=1463)
        0.02809942 = product of:
          0.05619884 = sum of:
            0.05619884 = weight(_text_:22 in 1463) [ClassicSimilarity], result of:
              0.05619884 = score(doc=1463,freq=2.0), product of:
                0.14525373 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.041479383 = queryNorm
                0.38690117 = fieldWeight in 1463, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1463)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Chronicles the early history of applying electronic computers to the task of translating natural languages, from the 1st suggestions by Warren Weaver in Mar 1947 to the 1st demonstration of a working, if limited, program in Jan 1954
    Date
    31. 7.1996 9:22:19
  7. Herrera-Viedma, E.; Cordón, O.; Herrera, J.C.; Luqe, M.: ¬An IRS based on multi-granular lnguistic information (2003) 0.04
    0.04170625 = product of:
      0.12511875 = sum of:
        0.09062983 = weight(_text_:21st in 2740) [ClassicSimilarity], result of:
          0.09062983 = score(doc=2740,freq=2.0), product of:
            0.2381352 = queryWeight, product of:
              5.74105 = idf(docFreq=385, maxDocs=44218)
              0.041479383 = queryNorm
            0.3805814 = fieldWeight in 2740, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.74105 = idf(docFreq=385, maxDocs=44218)
              0.046875 = fieldNorm(doc=2740)
        0.03448892 = product of:
          0.06897784 = sum of:
            0.06897784 = weight(_text_:century in 2740) [ClassicSimilarity], result of:
              0.06897784 = score(doc=2740,freq=2.0), product of:
                0.20775084 = queryWeight, product of:
                  5.0085325 = idf(docFreq=802, maxDocs=44218)
                  0.041479383 = queryNorm
                0.33202195 = fieldWeight in 2740, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.0085325 = idf(docFreq=802, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2740)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Source
    Challenges in knowledge representation and organization for the 21st century: Integration of knowledge across boundaries. Proceedings of the 7th ISKO International Conference Granada, Spain, July 10-13, 2002. Ed.: M. López-Huertas
  8. Martínez, F.; Martín, M.T.; Rivas, V.M.; Díaz, M.C.; Ureña, L.A.: Using neural networks for multiword recognition in IR (2003) 0.04
    0.04170625 = product of:
      0.12511875 = sum of:
        0.09062983 = weight(_text_:21st in 2777) [ClassicSimilarity], result of:
          0.09062983 = score(doc=2777,freq=2.0), product of:
            0.2381352 = queryWeight, product of:
              5.74105 = idf(docFreq=385, maxDocs=44218)
              0.041479383 = queryNorm
            0.3805814 = fieldWeight in 2777, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.74105 = idf(docFreq=385, maxDocs=44218)
              0.046875 = fieldNorm(doc=2777)
        0.03448892 = product of:
          0.06897784 = sum of:
            0.06897784 = weight(_text_:century in 2777) [ClassicSimilarity], result of:
              0.06897784 = score(doc=2777,freq=2.0), product of:
                0.20775084 = queryWeight, product of:
                  5.0085325 = idf(docFreq=802, maxDocs=44218)
                  0.041479383 = queryNorm
                0.33202195 = fieldWeight in 2777, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.0085325 = idf(docFreq=802, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2777)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Source
    Challenges in knowledge representation and organization for the 21st century: Integration of knowledge across boundaries. Proceedings of the 7th ISKO International Conference Granada, Spain, July 10-13, 2002. Ed.: M. López-Huertas
  9. Peis, E.; Herrera-Viedma, E.; Herrera, J.C.: On the evaluation of XML documents using Fuzzy linguistic techniques (2003) 0.04
    0.04170625 = product of:
      0.12511875 = sum of:
        0.09062983 = weight(_text_:21st in 2778) [ClassicSimilarity], result of:
          0.09062983 = score(doc=2778,freq=2.0), product of:
            0.2381352 = queryWeight, product of:
              5.74105 = idf(docFreq=385, maxDocs=44218)
              0.041479383 = queryNorm
            0.3805814 = fieldWeight in 2778, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.74105 = idf(docFreq=385, maxDocs=44218)
              0.046875 = fieldNorm(doc=2778)
        0.03448892 = product of:
          0.06897784 = sum of:
            0.06897784 = weight(_text_:century in 2778) [ClassicSimilarity], result of:
              0.06897784 = score(doc=2778,freq=2.0), product of:
                0.20775084 = queryWeight, product of:
                  5.0085325 = idf(docFreq=802, maxDocs=44218)
                  0.041479383 = queryNorm
                0.33202195 = fieldWeight in 2778, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.0085325 = idf(docFreq=802, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2778)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Source
    Challenges in knowledge representation and organization for the 21st century: Integration of knowledge across boundaries. Proceedings of the 7th ISKO International Conference Granada, Spain, July 10-13, 2002. Ed.: M. López-Huertas
  10. Paolillo, J.C.: Linguistics and the information sciences (2009) 0.03
    0.029697962 = product of:
      0.08909389 = sum of:
        0.069424294 = weight(_text_:history in 3840) [ClassicSimilarity], result of:
          0.069424294 = score(doc=3840,freq=2.0), product of:
            0.19296135 = queryWeight, product of:
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.041479383 = queryNorm
            0.3597834 = fieldWeight in 3840, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3840)
        0.019669592 = product of:
          0.039339185 = sum of:
            0.039339185 = weight(_text_:22 in 3840) [ClassicSimilarity], result of:
              0.039339185 = score(doc=3840,freq=2.0), product of:
                0.14525373 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.041479383 = queryNorm
                0.2708308 = fieldWeight in 3840, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3840)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Linguistics is the scientific study of language which emphasizes language spoken in everyday settings by human beings. It has a long history of interdisciplinarity, both internally and in contribution to other fields, including information science. A linguistic perspective is beneficial in many ways in information science, since it examines the relationship between the forms of meaningful expressions and their social, cognitive, institutional, and communicative context, these being two perspectives on information that are actively studied, to different degrees, in information science. Examples of issues relevant to information science are presented for which the approach taken under a linguistic perspective is illustrated.
    Date
    27. 8.2011 14:22:33
  11. From information to knowledge : conceptual and content analysis by computer (1995) 0.03
    0.026109848 = product of:
      0.07832954 = sum of:
        0.049588777 = weight(_text_:history in 5392) [ClassicSimilarity], result of:
          0.049588777 = score(doc=5392,freq=2.0), product of:
            0.19296135 = queryWeight, product of:
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.041479383 = queryNorm
            0.25698814 = fieldWeight in 5392, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5392)
        0.028740766 = product of:
          0.05748153 = sum of:
            0.05748153 = weight(_text_:century in 5392) [ClassicSimilarity], result of:
              0.05748153 = score(doc=5392,freq=2.0), product of:
                0.20775084 = queryWeight, product of:
                  5.0085325 = idf(docFreq=802, maxDocs=44218)
                  0.041479383 = queryNorm
                0.27668494 = fieldWeight in 5392, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.0085325 = idf(docFreq=802, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5392)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Content
    SCHMIDT, K.M.: Concepts - content - meaning: an introduction; DUCHASTEL, J. et al.: The SACAO project: using computation toward textual data analysis; PAQUIN, L.-C. u. L. DUPUY: An approach to expertise transfer: computer-assisted text analysis; HOGENRAAD, R., Y. BESTGEN u. J.-L. NYSTEN: Terrorist rhetoric: texture and architecture; MOHLER, P.P.: On the interaction between reading and computing: an interpretative approach to content analysis; LANCASHIRE, I.: Computer tools for cognitive stylistics; MERGENTHALER, E.: An outline of knowledge based text analysis; NAMENWIRTH, J.Z.: Ideography in computer-aided content analysis; WEBER, R.P. u. J.Z. Namenwirth: Content-analytic indicators: a self-critique; McKINNON, A.: Optimizing the aberrant frequency word technique; ROSATI, R.: Factor analysis in classical archaeology: export patterns of Attic pottery trade; PETRILLO, P.S.: Old and new worlds: ancient coinage and modern technology; DARANYI, S., S. MARJAI u.a.: Caryatids and the measurement of semiosis in architecture; ZARRI, G.P.: Intelligent information retrieval: an application in the field of historical biographical data; BOUCHARD, G., R. ROY u.a.: Computers and genealogy: from family reconstitution to population reconstruction; DEMÉLAS-BOHY, M.-D. u. M. RENAUD: Instability, networks and political parties: a political history expert system prototype; DARANYI, S., A. ABRANYI u. G. KOVACS: Knowledge extraction from ethnopoetic texts by multivariate statistical methods; FRAUTSCHI, R.L.: Measures of narrative voice in French prose fiction applied to textual samples from the enlightenment to the twentieth century; DANNENBERG, R. u.a.: A project in computer music: the musician's workbench
  12. Kocijan, K.: Visualizing natural language resources (2015) 0.03
    0.025174957 = product of:
      0.15104973 = sum of:
        0.15104973 = weight(_text_:21st in 2995) [ClassicSimilarity], result of:
          0.15104973 = score(doc=2995,freq=2.0), product of:
            0.2381352 = queryWeight, product of:
              5.74105 = idf(docFreq=385, maxDocs=44218)
              0.041479383 = queryNorm
            0.6343024 = fieldWeight in 2995, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.74105 = idf(docFreq=385, maxDocs=44218)
              0.078125 = fieldNorm(doc=2995)
      0.16666667 = coord(1/6)
    
    Source
    Re:inventing information science in the networked society: Proceedings of the 14th International Symposium on Information Science, Zadar/Croatia, 19th-21st May 2015. Eds.: F. Pehar, C. Schloegl u. C. Wolff
  13. Harari, Y.N.: ¬[Yuval-Noah-Harari-argues-that] AI has hacked the operating system of human civilisation (2023) 0.02
    0.016529594 = product of:
      0.099177554 = sum of:
        0.099177554 = weight(_text_:history in 953) [ClassicSimilarity], result of:
          0.099177554 = score(doc=953,freq=2.0), product of:
            0.19296135 = queryWeight, product of:
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.041479383 = queryNorm
            0.5139763 = fieldWeight in 953, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.078125 = fieldNorm(doc=953)
      0.16666667 = coord(1/6)
    
    Abstract
    Storytelling computers will change the course of human history, says the historian and philosopher.
  14. Pollitt, A.S.; Ellis, G.: Multilingual access to document databases (1993) 0.02
    0.015104972 = product of:
      0.09062983 = sum of:
        0.09062983 = weight(_text_:21st in 1302) [ClassicSimilarity], result of:
          0.09062983 = score(doc=1302,freq=2.0), product of:
            0.2381352 = queryWeight, product of:
              5.74105 = idf(docFreq=385, maxDocs=44218)
              0.041479383 = queryNorm
            0.3805814 = fieldWeight in 1302, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.74105 = idf(docFreq=385, maxDocs=44218)
              0.046875 = fieldNorm(doc=1302)
      0.16666667 = coord(1/6)
    
    Source
    Information as a Global Commodity - Communication, Processing and Use (CAIS/ACSI '93) : 21st Annual Conference Canadian Association for Information Science, Antigonish, Nova Scotia, Canada. July 1993
  15. Yang, C.C.; Luk, J.: Automatic generation of English/Chinese thesaurus based on a parallel corpus in laws (2003) 0.01
    0.014848981 = product of:
      0.044546943 = sum of:
        0.034712147 = weight(_text_:history in 1616) [ClassicSimilarity], result of:
          0.034712147 = score(doc=1616,freq=2.0), product of:
            0.19296135 = queryWeight, product of:
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.041479383 = queryNorm
            0.1798917 = fieldWeight in 1616, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1616)
        0.009834796 = product of:
          0.019669592 = sum of:
            0.019669592 = weight(_text_:22 in 1616) [ClassicSimilarity], result of:
              0.019669592 = score(doc=1616,freq=2.0), product of:
                0.14525373 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.041479383 = queryNorm
                0.1354154 = fieldWeight in 1616, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1616)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    The information available in languages other than English in the World Wide Web is increasing significantly. According to a report from Computer Economics in 1999, 54% of Internet users are English speakers ("English Will Dominate Web for Only Three More Years," Computer Economics, July 9, 1999, http://www.computereconomics. com/new4/pr/pr990610.html). However, it is predicted that there will be only 60% increase in Internet users among English speakers verses a 150% growth among nonEnglish speakers for the next five years. By 2005, 57% of Internet users will be non-English speakers. A report by CNN.com in 2000 showed that the number of Internet users in China had been increased from 8.9 million to 16.9 million from January to June in 2000 ("Report: China Internet users double to 17 million," CNN.com, July, 2000, http://cnn.org/2000/TECH/computing/07/27/ china.internet.reut/index.html). According to Nielsen/ NetRatings, there was a dramatic leap from 22.5 millions to 56.6 millions Internet users from 2001 to 2002. China had become the second largest global at-home Internet population in 2002 (US's Internet population was 166 millions) (Robyn Greenspan, "China Pulls Ahead of Japan," Internet.com, April 22, 2002, http://cyberatias.internet.com/big-picture/geographics/article/0,,5911_1013841,00. html). All of the evidences reveal the importance of crosslingual research to satisfy the needs in the near future. Digital library research has been focusing in structural and semantic interoperability in the past. Searching and retrieving objects across variations in protocols, formats and disciplines are widely explored (Schatz, B., & Chen, H. (1999). Digital libraries: technological advances and social impacts. IEEE Computer, Special Issue an Digital Libraries, February, 32(2), 45-50.; Chen, H., Yen, J., & Yang, C.C. (1999). International activities: development of Asian digital libraries. IEEE Computer, Special Issue an Digital Libraries, 32(2), 48-49.). However, research in crossing language boundaries, especially across European languages and Oriental languages, is still in the initial stage. In this proposal, we put our focus an cross-lingual semantic interoperability by developing automatic generation of a cross-lingual thesaurus based an English/Chinese parallel corpus. When the searchers encounter retrieval problems, Professional librarians usually consult the thesaurus to identify other relevant vocabularies. In the problem of searching across language boundaries, a cross-lingual thesaurus, which is generated by co-occurrence analysis and Hopfield network, can be used to generate additional semantically relevant terms that cannot be obtained from dictionary. In particular, the automatically generated cross-lingual thesaurus is able to capture the unknown words that do not exist in a dictionary, such as names of persons, organizations, and events. Due to Hong Kong's unique history background, both English and Chinese are used as official languages in all legal documents. Therefore, English/Chinese cross-lingual information retrieval is critical for applications in courts and the government. In this paper, we develop an automatic thesaurus by the Hopfield network based an a parallel corpus collected from the Web site of the Department of Justice of the Hong Kong Special Administrative Region (HKSAR) Government. Experiments are conducted to measure the precision and recall of the automatic generated English/Chinese thesaurus. The result Shows that such thesaurus is a promising tool to retrieve relevant terms, especially in the language that is not the same as the input term. The direct translation of the input term can also be retrieved in most of the cases.
  16. Comeau, D.C.; Wilbur, W.J.: Non-Word Identification or Spell Checking Without a Dictionary (2004) 0.01
    0.009917757 = product of:
      0.05950654 = sum of:
        0.05950654 = weight(_text_:history in 2092) [ClassicSimilarity], result of:
          0.05950654 = score(doc=2092,freq=2.0), product of:
            0.19296135 = queryWeight, product of:
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.041479383 = queryNorm
            0.3083858 = fieldWeight in 2092, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.046875 = fieldNorm(doc=2092)
      0.16666667 = coord(1/6)
    
    Abstract
    MEDLINE is a collection of more than 12 million references and abstracts covering recent life science literature. With its continued growth and cutting-edge terminology, spell-checking with a traditional lexicon based approach requires significant additional manual followup. In this work, an internal corpus based context quality rating a, frequency, and simple misspelling transformations are used to rank words from most likely to be misspellings to least likely. Eleven-point average precisions of 0.891 have been achieved within a class of 42,340 all alphabetic words having an a score less than 10. Our models predict that 16,274 or 38% of these words are misspellings. Based an test data, this result has a recall of 79% and a precision of 86%. In other words, spell checking can be done by statistics instead of with a dictionary. As an application we examine the time history of low a words in MEDLINE titles and abstracts.
  17. Sebastiani, F.: ¬A tutorial an automated text categorisation (1999) 0.01
    0.009917757 = product of:
      0.05950654 = sum of:
        0.05950654 = weight(_text_:history in 3390) [ClassicSimilarity], result of:
          0.05950654 = score(doc=3390,freq=2.0), product of:
            0.19296135 = queryWeight, product of:
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.041479383 = queryNorm
            0.3083858 = fieldWeight in 3390, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.046875 = fieldNorm(doc=3390)
      0.16666667 = coord(1/6)
    
    Abstract
    The automated categorisation (or classification) of texts into topical categories has a long history, dating back at least to 1960. Until the late '80s, the dominant approach to the problem involved knowledge-engineering automatic categorisers, i.e. manually building a set of rules encoding expert knowledge an how to classify documents. In the '90s, with the booming production and availability of on-line documents, automated text categorisation has witnessed an increased and renewed interest. A newer paradigm based an machine learning has superseded the previous approach. Within this paradigm, a general inductive process automatically builds a classifier by "learning", from a set of previously classified documents, the characteristics of one or more categories; the advantages are a very good effectiveness, a considerable savings in terms of expert manpower, and domain independence. In this tutorial we look at the main approaches that have been taken towards automatic text categorisation within the general machine learning paradigm. Issues of document indexing, classifier construction, and classifier evaluation, will be touched upon.
  18. Clark, M.; Kim, Y.; Kruschwitz, U.; Song, D.; Albakour, D.; Dignum, S.; Beresi, U.C.; Fasli, M.; Roeck, A De: Automatically structuring domain knowledge from text : an overview of current research (2012) 0.01
    0.009917757 = product of:
      0.05950654 = sum of:
        0.05950654 = weight(_text_:history in 2738) [ClassicSimilarity], result of:
          0.05950654 = score(doc=2738,freq=2.0), product of:
            0.19296135 = queryWeight, product of:
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.041479383 = queryNorm
            0.3083858 = fieldWeight in 2738, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.046875 = fieldNorm(doc=2738)
      0.16666667 = coord(1/6)
    
    Abstract
    This paper presents an overview of automatic methods for building domain knowledge structures (domain models) from text collections. Applications of domain models have a long history within knowledge engineering and artificial intelligence. In the last couple of decades they have surfaced noticeably as a useful tool within natural language processing, information retrieval and semantic web technology. Inspired by the ubiquitous propagation of domain model structures that are emerging in several research disciplines, we give an overview of the current research landscape and some techniques and approaches. We will also discuss trade-offs between different approaches and point to some recent trends.
  19. Lund, B.D.; Wang, T.; Mannuru, N.R.; Nie, B.; Shimray, S.; Wang, Z.: ChatGPT and a new academic reality : artificial Intelligence-written research papers and the ethics of the large language models in scholarly publishing (2023) 0.01
    0.009917757 = product of:
      0.05950654 = sum of:
        0.05950654 = weight(_text_:history in 943) [ClassicSimilarity], result of:
          0.05950654 = score(doc=943,freq=2.0), product of:
            0.19296135 = queryWeight, product of:
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.041479383 = queryNorm
            0.3083858 = fieldWeight in 943, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.046875 = fieldNorm(doc=943)
      0.16666667 = coord(1/6)
    
    Abstract
    This article discusses OpenAI's ChatGPT, a generative pre-trained transformer, which uses natural language processing to fulfill text-based user requests (i.e., a "chatbot"). The history and principles behind ChatGPT and similar models are discussed. This technology is then discussed in relation to its potential impact on academia and scholarly research and publishing. ChatGPT is seen as a potential model for the automated preparation of essays and other types of scholarly manuscripts. Potential ethical issues that could arise with the emergence of large language models like GPT-3, the underlying technology behind ChatGPT, and its usage by academics and researchers, are discussed and situated within the context of broader advancements in artificial intelligence, machine learning, and natural language processing for research and scholarly publishing.
  20. Bowker, L.; Ciro, J.B.: Machine translation and global research : towards improved machine translation literacy in the scholarly community (2019) 0.01
    0.00935055 = product of:
      0.0561033 = sum of:
        0.0561033 = weight(_text_:history in 5970) [ClassicSimilarity], result of:
          0.0561033 = score(doc=5970,freq=4.0), product of:
            0.19296135 = queryWeight, product of:
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.041479383 = queryNorm
            0.2907489 = fieldWeight in 5970, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.6519823 = idf(docFreq=1146, maxDocs=44218)
              0.03125 = fieldNorm(doc=5970)
      0.16666667 = coord(1/6)
    
    LCSH
    Literature / Translations / History and criticism
    Subject
    Literature / Translations / History and criticism

Years

Languages

  • e 52
  • d 16

Types

  • a 52
  • el 7
  • m 6
  • s 5
  • p 2
  • x 2
  • d 1
  • More… Less…