Search (751 results, page 1 of 38)

  • × theme_ss:"Computerlinguistik"
  1. Senez, D.: Developments in Systran (1995) 0.11
    0.11232709 = product of:
      0.22465418 = sum of:
        0.009566996 = weight(_text_:information in 8546) [ClassicSimilarity], result of:
          0.009566996 = score(doc=8546,freq=2.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.1551638 = fieldWeight in 8546, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=8546)
        0.018955056 = weight(_text_:for in 8546) [ClassicSimilarity], result of:
          0.018955056 = score(doc=8546,freq=6.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.28743884 = fieldWeight in 8546, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0625 = fieldNorm(doc=8546)
        0.01545607 = weight(_text_:the in 8546) [ClassicSimilarity], result of:
          0.01545607 = score(doc=8546,freq=8.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.27891195 = fieldWeight in 8546, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0625 = fieldNorm(doc=8546)
        0.15003714 = weight(_text_:commission in 8546) [ClassicSimilarity], result of:
          0.15003714 = score(doc=8546,freq=2.0), product of:
            0.24417253 = queryWeight, product of:
              6.9519553 = idf(docFreq=114, maxDocs=44218)
              0.035122856 = queryNorm
            0.61447185 = fieldWeight in 8546, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.9519553 = idf(docFreq=114, maxDocs=44218)
              0.0625 = fieldNorm(doc=8546)
        0.015182858 = weight(_text_:of in 8546) [ClassicSimilarity], result of:
          0.015182858 = score(doc=8546,freq=8.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.27643585 = fieldWeight in 8546, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=8546)
        0.01545607 = weight(_text_:the in 8546) [ClassicSimilarity], result of:
          0.01545607 = score(doc=8546,freq=8.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.27891195 = fieldWeight in 8546, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0625 = fieldNorm(doc=8546)
      0.5 = coord(6/12)
    
    Abstract
    Systran, the European Commission's multilingual machine translation system, is a fast service which is available to all Commission officials. The computer cannot match the skills of the professional translator, who must continue to be responsible for all texts which are legally binding or which are for publication. But machine translation can deal, in a matter of minutes, with short-lived documents, designed, say, for information or preparatory work, and which are required urgently. It can also give a broad view of a paper in an unfamiliar language, so that an official can decide how much, if any, of it needs to go to translators
  2. Wenzel, F.: Semantische Eingrenzung im Freitext-Retrieval auf der Basis morphologischer Segmentierungen (1980) 0.09
    0.08690446 = product of:
      0.17380892 = sum of:
        0.020549515 = product of:
          0.061648544 = sum of:
            0.061648544 = weight(_text_:f in 2037) [ClassicSimilarity], result of:
              0.061648544 = score(doc=2037,freq=2.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.4403713 = fieldWeight in 2037, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2037)
          0.33333334 = coord(1/3)
        0.011958744 = weight(_text_:information in 2037) [ClassicSimilarity], result of:
          0.011958744 = score(doc=2037,freq=2.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.19395474 = fieldWeight in 2037, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=2037)
        0.08467974 = weight(_text_:dokumentation in 2037) [ClassicSimilarity], result of:
          0.08467974 = score(doc=2037,freq=2.0), product of:
            0.16407113 = queryWeight, product of:
              4.671349 = idf(docFreq=1124, maxDocs=44218)
              0.035122856 = queryNorm
            0.516116 = fieldWeight in 2037, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.671349 = idf(docFreq=1124, maxDocs=44218)
              0.078125 = fieldNorm(doc=2037)
        0.021600515 = weight(_text_:the in 2037) [ClassicSimilarity], result of:
          0.021600515 = score(doc=2037,freq=10.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3897913 = fieldWeight in 2037, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.078125 = fieldNorm(doc=2037)
        0.013419878 = weight(_text_:of in 2037) [ClassicSimilarity], result of:
          0.013419878 = score(doc=2037,freq=4.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.24433708 = fieldWeight in 2037, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.078125 = fieldNorm(doc=2037)
        0.021600515 = weight(_text_:the in 2037) [ClassicSimilarity], result of:
          0.021600515 = score(doc=2037,freq=10.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3897913 = fieldWeight in 2037, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.078125 = fieldNorm(doc=2037)
      0.5 = coord(6/12)
    
    Abstract
    The basic problem in freetext retrieval is that the retrieval language is not properly adapted to that of the author. Morphological segmentation, where words with the same root are grouped together in the inverted file, is a good eliminator of noise and information loss, providing high recall but low precision
    Source
    Nachrichten für Dokumentation. 31(1980) H.1, S.29-35
  3. Lehrndorfer, A.: Kontrolliertes Deutsch : Linguistische und sprachpsychologische Leitlinien für eine (maschinell) kontrollierte Sprache in der Technischen Dokumentation (1996) 0.07
    0.07005926 = product of:
      0.14011852 = sum of:
        0.015249942 = weight(_text_:und in 2039) [ClassicSimilarity], result of:
          0.015249942 = score(doc=2039,freq=2.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.19590102 = fieldWeight in 2039, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=2039)
        0.06774379 = weight(_text_:dokumentation in 2039) [ClassicSimilarity], result of:
          0.06774379 = score(doc=2039,freq=2.0), product of:
            0.16407113 = queryWeight, product of:
              4.671349 = idf(docFreq=1124, maxDocs=44218)
              0.035122856 = queryNorm
            0.41289282 = fieldWeight in 2039, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.671349 = idf(docFreq=1124, maxDocs=44218)
              0.0625 = fieldNorm(doc=2039)
        0.015476737 = weight(_text_:for in 2039) [ClassicSimilarity], result of:
          0.015476737 = score(doc=2039,freq=4.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.23469281 = fieldWeight in 2039, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0625 = fieldNorm(doc=2039)
        0.01545607 = weight(_text_:the in 2039) [ClassicSimilarity], result of:
          0.01545607 = score(doc=2039,freq=8.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.27891195 = fieldWeight in 2039, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0625 = fieldNorm(doc=2039)
        0.010735902 = weight(_text_:of in 2039) [ClassicSimilarity], result of:
          0.010735902 = score(doc=2039,freq=4.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.19546966 = fieldWeight in 2039, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=2039)
        0.01545607 = weight(_text_:the in 2039) [ClassicSimilarity], result of:
          0.01545607 = score(doc=2039,freq=8.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.27891195 = fieldWeight in 2039, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0625 = fieldNorm(doc=2039)
      0.5 = coord(6/12)
    
    Content
    The book comprises 8 chapters and 5 appendices: Artificial languages, sublanguages and controlled languages; Theory for language planning, especially for planning of controlled languages; The language situation of technical documentation in Germany; The Lexicon; The syntax; Perspectives; Literature
  4. Schneider, J.W.; Borlund, P.: ¬A bibliometric-based semiautomatic approach to identification of candidate thesaurus terms : parsing and filtering of noun phrases from citation contexts (2005) 0.06
    0.062578455 = product of:
      0.107277356 = sum of:
        0.01438466 = product of:
          0.04315398 = sum of:
            0.04315398 = weight(_text_:f in 156) [ClassicSimilarity], result of:
              0.04315398 = score(doc=156,freq=2.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.3082599 = fieldWeight in 156, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=156)
          0.33333334 = coord(1/3)
        0.008371122 = weight(_text_:information in 156) [ClassicSimilarity], result of:
          0.008371122 = score(doc=156,freq=2.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.13576832 = fieldWeight in 156, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=156)
        0.013343699 = weight(_text_:und in 156) [ClassicSimilarity], result of:
          0.013343699 = score(doc=156,freq=2.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.17141339 = fieldWeight in 156, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=156)
        0.01912591 = weight(_text_:the in 156) [ClassicSimilarity], result of:
          0.01912591 = score(doc=156,freq=16.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.34513593 = fieldWeight in 156, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0546875 = fieldNorm(doc=156)
        0.016270736 = weight(_text_:of in 156) [ClassicSimilarity], result of:
          0.016270736 = score(doc=156,freq=12.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.29624295 = fieldWeight in 156, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=156)
        0.01912591 = weight(_text_:the in 156) [ClassicSimilarity], result of:
          0.01912591 = score(doc=156,freq=16.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.34513593 = fieldWeight in 156, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0546875 = fieldNorm(doc=156)
        0.016655317 = product of:
          0.033310633 = sum of:
            0.033310633 = weight(_text_:22 in 156) [ClassicSimilarity], result of:
              0.033310633 = score(doc=156,freq=2.0), product of:
                0.12299426 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035122856 = queryNorm
                0.2708308 = fieldWeight in 156, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=156)
          0.5 = coord(1/2)
      0.5833333 = coord(7/12)
    
    Abstract
    The present study investigates the ability of a bibliometric based semi-automatic method to select candidate thesaurus terms from citation contexts. The method consists of document co-citation analysis, citation context analysis, and noun phrase parsing. The investigation is carried out within the specialty area of periodontology. The results clearly demonstrate that the method is able to select important candidate thesaurus terms within the chosen specialty area.
    Date
    8. 3.2007 19:55:22
    Source
    Context: nature, impact and role. 5th International Conference an Conceptions of Library and Information Sciences, CoLIS 2005 Glasgow, UK, June 2005. Ed. by F. Crestani u. I. Ruthven
    Theme
    Konzeption und Anwendung des Prinzips Thesaurus
  5. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.06
    0.060159884 = product of:
      0.12031977 = sum of:
        0.055784404 = product of:
          0.16735321 = sum of:
            0.16735321 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
              0.16735321 = score(doc=562,freq=2.0), product of:
                0.29777196 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.035122856 = queryNorm
                0.56201804 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.33333334 = coord(1/3)
        0.011607553 = weight(_text_:for in 562) [ClassicSimilarity], result of:
          0.011607553 = score(doc=562,freq=4.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.17601961 = fieldWeight in 562, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.01296031 = weight(_text_:the in 562) [ClassicSimilarity], result of:
          0.01296031 = score(doc=562,freq=10.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.2338748 = fieldWeight in 562, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.012731214 = weight(_text_:of in 562) [ClassicSimilarity], result of:
          0.012731214 = score(doc=562,freq=10.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.23179851 = fieldWeight in 562, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.01296031 = weight(_text_:the in 562) [ClassicSimilarity], result of:
          0.01296031 = score(doc=562,freq=10.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.2338748 = fieldWeight in 562, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.014275986 = product of:
          0.028551972 = sum of:
            0.028551972 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
              0.028551972 = score(doc=562,freq=2.0), product of:
                0.12299426 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035122856 = queryNorm
                0.23214069 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.5 = coord(1/2)
      0.5 = coord(6/12)
    
    Abstract
    Document representations for text classification are typically based on the classical Bag-Of-Words paradigm. This approach comes with deficiencies that motivate the integration of features on a higher semantic level than single words. In this paper we propose an enhancement of the classical document representation through concepts extracted from background knowledge. Boosting is used for actual classification. Experimental evaluations on two well known text corpora support our approach through consistent improvement of the results.
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
    Source
    Proceedings of the 4th IEEE International Conference on Data Mining (ICDM 2004), 1-4 November 2004, Brighton, UK
  6. Schmitz, K.-D.: Projektforschung und Infrastrukturen im Bereich der Terminologie : Wie kann die Wirtschaft davon profitieren? (2000) 0.06
    0.059518293 = product of:
      0.119036585 = sum of:
        0.037933756 = weight(_text_:und in 5568) [ClassicSimilarity], result of:
          0.037933756 = score(doc=5568,freq=22.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.48729765 = fieldWeight in 5568, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=5568)
        0.05080784 = weight(_text_:dokumentation in 5568) [ClassicSimilarity], result of:
          0.05080784 = score(doc=5568,freq=2.0), product of:
            0.16407113 = queryWeight, product of:
              4.671349 = idf(docFreq=1124, maxDocs=44218)
              0.035122856 = queryNorm
            0.3096696 = fieldWeight in 5568, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.671349 = idf(docFreq=1124, maxDocs=44218)
              0.046875 = fieldNorm(doc=5568)
        0.008207779 = weight(_text_:for in 5568) [ClassicSimilarity], result of:
          0.008207779 = score(doc=5568,freq=2.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.12446466 = fieldWeight in 5568, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=5568)
        0.008196819 = weight(_text_:the in 5568) [ClassicSimilarity], result of:
          0.008196819 = score(doc=5568,freq=4.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.1479154 = fieldWeight in 5568, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=5568)
        0.005693572 = weight(_text_:of in 5568) [ClassicSimilarity], result of:
          0.005693572 = score(doc=5568,freq=2.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.103663445 = fieldWeight in 5568, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=5568)
        0.008196819 = weight(_text_:the in 5568) [ClassicSimilarity], result of:
          0.008196819 = score(doc=5568,freq=4.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.1479154 = fieldWeight in 5568, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=5568)
      0.5 = coord(6/12)
    
    Abstract
    In der heutigen Informationsgesellschaft bieten sich der Industrie neue Perspektiven für Kommunikation und Handel auf dem europäischen und internationalen Markt; beide Märkte sind von einer großen sprachlichen, kulturellen und gesellschaftlichen Vielfalt geprägt. Uni Nutzen aus diesen neuen Möglichkeiten zu ziehen und um weiterhin konkurrenzfähig zu bleiben, muß die Industrie spezifische und adäquate Lösungen zur Überwindung der Sprachbarrieren finden. Voraussetzung hierfür ist die genaue Definition, systematische Ordnung und exakte Benennung der Begriffe innerhalb der jeweiligen Fachgebiete, in der eigenen Sprache ebenso wie in den Fremdsprachen. Genau dies sind die Themenbereiche, mit dem sich die Terminologiewissenschaft und die praktische Temninologiearbeit beschäftigen. Die Ergebnisse der Terminologiearbeit im Unternehmen beeinflussen Konstruktion, Produktion, Einkauf, Marketing und Verkauf, Vertragswesen, technische Dokumentation und Übersetzung
    Source
    Sprachtechnologie für eine dynamische Wirtschaft im Medienzeitalter - Language technologies for dynamic business in the age of the media - L'ingénierie linguistique au service de la dynamisation économique à l'ère du multimédia: Tagungsakten der XXVI. Jahrestagung der Internationalen Vereinigung Sprache und Wirtschaft e.V., 23.-25.11.2000, Fachhochschule Köln. Hrsg.: K.-D. Schmitz
  7. Dietze, J.; Völkel, H.: Verifikation einer Methode der lexikalischen Semantik : zur computergestützten Bestimmung der semantischen Konsistenz und des semantischen Abstands (1992) 0.06
    0.057428405 = product of:
      0.13782817 = sum of:
        0.015249942 = weight(_text_:und in 6680) [ClassicSimilarity], result of:
          0.015249942 = score(doc=6680,freq=2.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.19590102 = fieldWeight in 6680, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0625 = fieldNorm(doc=6680)
        0.06774379 = weight(_text_:dokumentation in 6680) [ClassicSimilarity], result of:
          0.06774379 = score(doc=6680,freq=2.0), product of:
            0.16407113 = queryWeight, product of:
              4.671349 = idf(docFreq=1124, maxDocs=44218)
              0.035122856 = queryNorm
            0.41289282 = fieldWeight in 6680, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.671349 = idf(docFreq=1124, maxDocs=44218)
              0.0625 = fieldNorm(doc=6680)
        0.018929742 = weight(_text_:the in 6680) [ClassicSimilarity], result of:
          0.018929742 = score(doc=6680,freq=12.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.34159598 = fieldWeight in 6680, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0625 = fieldNorm(doc=6680)
        0.016974952 = weight(_text_:of in 6680) [ClassicSimilarity], result of:
          0.016974952 = score(doc=6680,freq=10.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.3090647 = fieldWeight in 6680, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=6680)
        0.018929742 = weight(_text_:the in 6680) [ClassicSimilarity], result of:
          0.018929742 = score(doc=6680,freq=12.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.34159598 = fieldWeight in 6680, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0625 = fieldNorm(doc=6680)
      0.41666666 = coord(5/12)
    
    Abstract
    Uses a semantic field 'linguistic communication' of 735 verbs to verify two numerically based methods working with the semic cooccurrence interval due to the semic micro-structure of a lexeme. The weak point of this procedure is the one-stage classification of the semantic features (semes) of the field
    Source
    Nachrichten für Dokumentation. 43(1992) H.3, S.159-164
  8. Rosemblat, G.; Resnick, M.P.; Auston, I.; Shin, D.; Sneiderman, C.; Fizsman, M.; Rindflesch, T.C.: Extending SemRep to the public health domain (2013) 0.06
    0.05681357 = product of:
      0.11362714 = sum of:
        0.016044341 = weight(_text_:information in 2096) [ClassicSimilarity], result of:
          0.016044341 = score(doc=2096,freq=10.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.2602176 = fieldWeight in 2096, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2096)
        0.014216291 = weight(_text_:for in 2096) [ClassicSimilarity], result of:
          0.014216291 = score(doc=2096,freq=6.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.21557912 = fieldWeight in 2096, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=2096)
        0.018328644 = weight(_text_:the in 2096) [ClassicSimilarity], result of:
          0.018328644 = score(doc=2096,freq=20.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3307489 = fieldWeight in 2096, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=2096)
        0.013946345 = weight(_text_:of in 2096) [ClassicSimilarity], result of:
          0.013946345 = score(doc=2096,freq=12.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.25392252 = fieldWeight in 2096, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=2096)
        0.018328644 = weight(_text_:the in 2096) [ClassicSimilarity], result of:
          0.018328644 = score(doc=2096,freq=20.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3307489 = fieldWeight in 2096, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=2096)
        0.032762878 = product of:
          0.065525755 = sum of:
            0.065525755 = weight(_text_:communities in 2096) [ClassicSimilarity], result of:
              0.065525755 = score(doc=2096,freq=2.0), product of:
                0.18632571 = queryWeight, product of:
                  5.3049703 = idf(docFreq=596, maxDocs=44218)
                  0.035122856 = queryNorm
                0.35167316 = fieldWeight in 2096, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.3049703 = idf(docFreq=596, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2096)
          0.5 = coord(1/2)
      0.5 = coord(6/12)
    
    Abstract
    We describe the use of a domain-independent method to extend a natural language processing (NLP) application, SemRep (Rindflesch, Fiszman, & Libbus, 2005), based on the knowledge sources afforded by the Unified Medical Language System (UMLS®; Humphreys, Lindberg, Schoolman, & Barnett, 1998) to support the area of health promotion within the public health domain. Public health professionals require good information about successful health promotion policies and programs that might be considered for application within their own communities. Our effort seeks to improve access to relevant information for the public health profession, to help those in the field remain an information-savvy workforce. Natural language processing and semantic techniques hold promise to help public health professionals navigate the growing ocean of information by organizing and structuring this knowledge into a focused public health framework paired with a user-friendly visualization application as a way to summarize results of PubMed® searches in this field of knowledge.
    Source
    Journal of the American Society for Information Science and Technology. 64(2013) no.10, S.1963-1974
  9. Ahmad, F.; Yusoff, M.; Sembok, T.M.T.: Experiments with a stemming algorithm for Malay words (1996) 0.06
    0.055686876 = product of:
      0.11137375 = sum of:
        0.016439613 = product of:
          0.049318835 = sum of:
            0.049318835 = weight(_text_:f in 6504) [ClassicSimilarity], result of:
              0.049318835 = score(doc=6504,freq=2.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.35229704 = fieldWeight in 6504, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6504)
          0.33333334 = coord(1/3)
        0.013529775 = weight(_text_:information in 6504) [ClassicSimilarity], result of:
          0.013529775 = score(doc=6504,freq=4.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.21943474 = fieldWeight in 6504, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=6504)
        0.02188741 = weight(_text_:for in 6504) [ClassicSimilarity], result of:
          0.02188741 = score(doc=6504,freq=8.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.33190575 = fieldWeight in 6504, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0625 = fieldNorm(doc=6504)
        0.023184106 = weight(_text_:the in 6504) [ClassicSimilarity], result of:
          0.023184106 = score(doc=6504,freq=18.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.41836792 = fieldWeight in 6504, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0625 = fieldNorm(doc=6504)
        0.013148742 = weight(_text_:of in 6504) [ClassicSimilarity], result of:
          0.013148742 = score(doc=6504,freq=6.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.23940048 = fieldWeight in 6504, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=6504)
        0.023184106 = weight(_text_:the in 6504) [ClassicSimilarity], result of:
          0.023184106 = score(doc=6504,freq=18.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.41836792 = fieldWeight in 6504, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0625 = fieldNorm(doc=6504)
      0.5 = coord(6/12)
    
    Abstract
    Stemming is used in information retrieval systems to reduce variant word forms to common roots in order to improve retrieval effectiveness. As in other languages, there is a need for an effective stemming algorithm for the indexing and retrieval of Malay documents. The Malay stemming algorithm developed by Othman is studied and new versions proposed to enhance its performance. The improvements relate to the order in which the dictionary id looked-up, the order in which the morphological rules are applied, and the number of rules
    Source
    Journal of the American Society for Information Science. 47(1996) no.12, S.909-918
  10. Luo, L.; Ju, J.; Li, Y.-F.; Haffari, G.; Xiong, B.; Pan, S.: ChatRule: mining logical rules with large language models for knowledge graph reasoning (2023) 0.05
    0.054730415 = product of:
      0.093823574 = sum of:
        0.010274758 = product of:
          0.030824272 = sum of:
            0.030824272 = weight(_text_:f in 1171) [ClassicSimilarity], result of:
              0.030824272 = score(doc=1171,freq=2.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.22018565 = fieldWeight in 1171, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1171)
          0.33333334 = coord(1/3)
        0.005979372 = weight(_text_:information in 1171) [ClassicSimilarity], result of:
          0.005979372 = score(doc=1171,freq=2.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.09697737 = fieldWeight in 1171, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1171)
        0.015294294 = weight(_text_:for in 1171) [ClassicSimilarity], result of:
          0.015294294 = score(doc=1171,freq=10.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.2319262 = fieldWeight in 1171, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1171)
        0.018072287 = weight(_text_:the in 1171) [ClassicSimilarity], result of:
          0.018072287 = score(doc=1171,freq=28.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3261228 = fieldWeight in 1171, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1171)
        0.014233928 = weight(_text_:of in 1171) [ClassicSimilarity], result of:
          0.014233928 = score(doc=1171,freq=18.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.25915858 = fieldWeight in 1171, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1171)
        0.018072287 = weight(_text_:the in 1171) [ClassicSimilarity], result of:
          0.018072287 = score(doc=1171,freq=28.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3261228 = fieldWeight in 1171, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1171)
        0.011896656 = product of:
          0.023793312 = sum of:
            0.023793312 = weight(_text_:22 in 1171) [ClassicSimilarity], result of:
              0.023793312 = score(doc=1171,freq=2.0), product of:
                0.12299426 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035122856 = queryNorm
                0.19345059 = fieldWeight in 1171, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1171)
          0.5 = coord(1/2)
      0.5833333 = coord(7/12)
    
    Abstract
    Logical rules are essential for uncovering the logical connections between relations, which could improve the reasoning performance and provide interpretable results on knowledge graphs (KGs). Although there have been many efforts to mine meaningful logical rules over KGs, existing methods suffer from the computationally intensive searches over the rule space and a lack of scalability for large-scale KGs. Besides, they often ignore the semantics of relations which is crucial for uncovering logical connections. Recently, large language models (LLMs) have shown impressive performance in the field of natural language processing and various applications, owing to their emergent ability and generalizability. In this paper, we propose a novel framework, ChatRule, unleashing the power of large language models for mining logical rules over knowledge graphs. Specifically, the framework is initiated with an LLM-based rule generator, leveraging both the semantic and structural information of KGs to prompt LLMs to generate logical rules. To refine the generated rules, a rule ranking module estimates the rule quality by incorporating facts from existing KGs. Last, a rule validator harnesses the reasoning ability of LLMs to validate the logical correctness of ranked rules through chain-of-thought reasoning. ChatRule is evaluated on four large-scale KGs, w.r.t. different rule quality metrics and downstream tasks, showing the effectiveness and scalability of our method.
    Date
    23.11.2023 19:07:22
  11. Wanner, L.: Lexical choice in text generation and machine translation (1996) 0.05
    0.05197804 = product of:
      0.10395608 = sum of:
        0.009566996 = weight(_text_:information in 8521) [ClassicSimilarity], result of:
          0.009566996 = score(doc=8521,freq=2.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.1551638 = fieldWeight in 8521, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=8521)
        0.010943705 = weight(_text_:for in 8521) [ClassicSimilarity], result of:
          0.010943705 = score(doc=8521,freq=2.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.16595288 = fieldWeight in 8521, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0625 = fieldNorm(doc=8521)
        0.025630994 = weight(_text_:the in 8521) [ClassicSimilarity], result of:
          0.025630994 = score(doc=8521,freq=22.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.46252316 = fieldWeight in 8521, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0625 = fieldNorm(doc=8521)
        0.013148742 = weight(_text_:of in 8521) [ClassicSimilarity], result of:
          0.013148742 = score(doc=8521,freq=6.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.23940048 = fieldWeight in 8521, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=8521)
        0.025630994 = weight(_text_:the in 8521) [ClassicSimilarity], result of:
          0.025630994 = score(doc=8521,freq=22.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.46252316 = fieldWeight in 8521, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0625 = fieldNorm(doc=8521)
        0.019034648 = product of:
          0.038069297 = sum of:
            0.038069297 = weight(_text_:22 in 8521) [ClassicSimilarity], result of:
              0.038069297 = score(doc=8521,freq=2.0), product of:
                0.12299426 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035122856 = queryNorm
                0.30952093 = fieldWeight in 8521, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=8521)
          0.5 = coord(1/2)
      0.5 = coord(6/12)
    
    Abstract
    Presents the state of the art in lexical choice research in text generation and machine translation. Discusses the existing implementations with respect to: the place of lexical choice in the overall generation rates; the information flow within the generation process and the consequences thereof for lexical choice; the internal organization of the lexical choice process; and the phenomena covered by lexical choice. Identifies possible future directions in lexical choice research
    Date
    31. 7.1996 9:22:19
  12. Betrand-Gastaldy, S.: ¬La modelisation de l'analyse documentaire : à la convergence de la semiotique, de la psychologie cognitive et de l'intelligence (1995) 0.05
    0.051630914 = product of:
      0.10326183 = sum of:
        0.012329709 = product of:
          0.036989126 = sum of:
            0.036989126 = weight(_text_:f in 5377) [ClassicSimilarity], result of:
              0.036989126 = score(doc=5377,freq=2.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.26422277 = fieldWeight in 5377, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5377)
          0.33333334 = coord(1/3)
        0.012427893 = weight(_text_:information in 5377) [ClassicSimilarity], result of:
          0.012427893 = score(doc=5377,freq=6.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.20156369 = fieldWeight in 5377, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=5377)
        0.011607553 = weight(_text_:for in 5377) [ClassicSimilarity], result of:
          0.011607553 = score(doc=5377,freq=4.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.17601961 = fieldWeight in 5377, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=5377)
        0.023184106 = weight(_text_:the in 5377) [ClassicSimilarity], result of:
          0.023184106 = score(doc=5377,freq=32.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.41836792 = fieldWeight in 5377, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=5377)
        0.020528466 = weight(_text_:of in 5377) [ClassicSimilarity], result of:
          0.020528466 = score(doc=5377,freq=26.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.37376386 = fieldWeight in 5377, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=5377)
        0.023184106 = weight(_text_:the in 5377) [ClassicSimilarity], result of:
          0.023184106 = score(doc=5377,freq=32.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.41836792 = fieldWeight in 5377, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=5377)
      0.5 = coord(6/12)
    
    Abstract
    Textual semiotics and cognitive psychology are advocated to model several types of documentary analysis. Proposes a theoretical model which combines elements from the 2 disciplines. Thanks to the addition of values of properties pertaining to different semiotic systems to the primary and secondary texts, one can retrieve the units and the characteristics valued by a group of indexers or by one individual. The cognitive studies of the experts confirm or complete the textual analysis. Examples from the findings obtained by the statistic-linguistic analysis of 2 corpora illustrate the usefulness of the methodology, especially for the conception of expert systems to assist whatever kind of reading
    Imprint
    Alberta : Alberta University, School of Library and Information Studies
    Language
    f
    Source
    Connectedness: information, systems, people, organizations. Proceedings of CAIS/ACSI 95, the proceedings of the 23rd Annual Conference of the Canadian Association for Information Science. Ed. by Hope A. Olson and Denis B. Ward
  13. Seelbach, D.: Computerlinguistik und Dokumentation : keyphrases in Dokumentationsprozessen (1975) 0.05
    0.050927084 = product of:
      0.20370834 = sum of:
        0.010147331 = weight(_text_:information in 299) [ClassicSimilarity], result of:
          0.010147331 = score(doc=299,freq=4.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.16457605 = fieldWeight in 299, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=299)
        0.049854714 = weight(_text_:und in 299) [ClassicSimilarity], result of:
          0.049854714 = score(doc=299,freq=38.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.6404345 = fieldWeight in 299, product of:
              6.164414 = tf(freq=38.0), with freq of:
                38.0 = termFreq=38.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=299)
        0.14370629 = weight(_text_:dokumentation in 299) [ClassicSimilarity], result of:
          0.14370629 = score(doc=299,freq=16.0), product of:
            0.16407113 = queryWeight, product of:
              4.671349 = idf(docFreq=1124, maxDocs=44218)
              0.035122856 = queryNorm
            0.875878 = fieldWeight in 299, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              4.671349 = idf(docFreq=1124, maxDocs=44218)
              0.046875 = fieldNorm(doc=299)
      0.25 = coord(3/12)
    
    Classification
    ES 950 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Maschinelle Redeanlyse
    ES 955 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Maschinelle Referatherstellung, linguistische Dokumentation und Information
    Imprint
    München : Verlag Dokumentation
    RSWK
    Dokumentation (BVB)
    Linguistische Datenverarbeitung / Dokumentation (BVB)
    RVK
    ES 950 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Maschinelle Redeanlyse
    ES 955 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Maschinelle Referatherstellung, linguistische Dokumentation und Information
    Subject
    Dokumentation (BVB)
    Linguistische Datenverarbeitung / Dokumentation (BVB)
  14. Liddy, E.D.: Natural language processing for information retrieval and knowledge discovery (1998) 0.05
    0.050493967 = product of:
      0.10098793 = sum of:
        0.018718397 = weight(_text_:information in 2345) [ClassicSimilarity], result of:
          0.018718397 = score(doc=2345,freq=10.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.3035872 = fieldWeight in 2345, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2345)
        0.016585674 = weight(_text_:for in 2345) [ClassicSimilarity], result of:
          0.016585674 = score(doc=2345,freq=6.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.25150898 = fieldWeight in 2345, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2345)
        0.01512036 = weight(_text_:the in 2345) [ClassicSimilarity], result of:
          0.01512036 = score(doc=2345,freq=10.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.2728539 = fieldWeight in 2345, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2345)
        0.01878783 = weight(_text_:of in 2345) [ClassicSimilarity], result of:
          0.01878783 = score(doc=2345,freq=16.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.34207192 = fieldWeight in 2345, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2345)
        0.01512036 = weight(_text_:the in 2345) [ClassicSimilarity], result of:
          0.01512036 = score(doc=2345,freq=10.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.2728539 = fieldWeight in 2345, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2345)
        0.016655317 = product of:
          0.033310633 = sum of:
            0.033310633 = weight(_text_:22 in 2345) [ClassicSimilarity], result of:
              0.033310633 = score(doc=2345,freq=2.0), product of:
                0.12299426 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035122856 = queryNorm
                0.2708308 = fieldWeight in 2345, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2345)
          0.5 = coord(1/2)
      0.5 = coord(6/12)
    
    Abstract
    Natural language processing (NLP) is a powerful technology for the vital tasks of information retrieval (IR) and knowledge discovery (KD) which, in turn, feed the visualization systems of the present and future and enable knowledge workers to focus more of their time on the vital tasks of analysis and prediction
    Date
    22. 9.1997 19:16:05
    Imprint
    Urbana-Champaign, IL : Illinois University at Urbana-Champaign, Graduate School of Library and Information Science
    Source
    Visualizing subject access for 21st century information resources: Papers presented at the 1997 Clinic on Library Applications of Data Processing, 2-4 Mar 1997, Graduate School of Library and Information Science, University of Illinois at Urbana-Champaign. Ed.: P.A. Cochrane et al
  15. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.05
    0.05011634 = product of:
      0.12027922 = sum of:
        0.055784404 = product of:
          0.16735321 = sum of:
            0.16735321 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
              0.16735321 = score(doc=862,freq=2.0), product of:
                0.29777196 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.035122856 = queryNorm
                0.56201804 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=862)
          0.33333334 = coord(1/3)
        0.011607553 = weight(_text_:for in 862) [ClassicSimilarity], result of:
          0.011607553 = score(doc=862,freq=4.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.17601961 = fieldWeight in 862, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
        0.020078024 = weight(_text_:the in 862) [ClassicSimilarity], result of:
          0.020078024 = score(doc=862,freq=24.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.36231726 = fieldWeight in 862, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
        0.012731214 = weight(_text_:of in 862) [ClassicSimilarity], result of:
          0.012731214 = score(doc=862,freq=10.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.23179851 = fieldWeight in 862, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
        0.020078024 = weight(_text_:the in 862) [ClassicSimilarity], result of:
          0.020078024 = score(doc=862,freq=24.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.36231726 = fieldWeight in 862, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
      0.41666666 = coord(5/12)
    
    Abstract
    This research revisits the classic Turing test and compares recent large language models such as ChatGPT for their abilities to reproduce human-level comprehension and compelling text generation. Two task challenges- summary and question answering- prompt ChatGPT to produce original content (98-99%) from a single text entry and sequential questions initially posed by Turing in 1950. We score the original and generated content against the OpenAI GPT-2 Output Detector from 2019, and establish multiple cases where the generated content proves original and undetectable (98%). The question of a machine fooling a human judge recedes in this work relative to the question of "how would one prove it?" The original contribution of the work presents a metric and simple grammatical set for understanding the writing mechanics of chatbots in evaluating their readability and statistical clarity, engagement, delivery, overall quality, and plagiarism risks. While Turing's original prose scores at least 14% below the machine-generated output, whether an algorithm displays hints of Turing's true initial thoughts (the "Lovelace 2.0" test) remains unanswerable.
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
  16. Galvez, C.; Moya-Anegón, F. de; Solana, V.H.: Term conflation methods in information retrieval : non-linguistic and linguistic approaches (2005) 0.05
    0.049894966 = product of:
      0.09978993 = sum of:
        0.012329709 = product of:
          0.036989126 = sum of:
            0.036989126 = weight(_text_:f in 4394) [ClassicSimilarity], result of:
              0.036989126 = score(doc=4394,freq=2.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.26422277 = fieldWeight in 4394, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4394)
          0.33333334 = coord(1/3)
        0.010147331 = weight(_text_:information in 4394) [ClassicSimilarity], result of:
          0.010147331 = score(doc=4394,freq=4.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.16457605 = fieldWeight in 4394, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=4394)
        0.014216291 = weight(_text_:for in 4394) [ClassicSimilarity], result of:
          0.014216291 = score(doc=4394,freq=6.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.21557912 = fieldWeight in 4394, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=4394)
        0.021686744 = weight(_text_:the in 4394) [ClassicSimilarity], result of:
          0.021686744 = score(doc=4394,freq=28.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.39134735 = fieldWeight in 4394, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=4394)
        0.019723112 = weight(_text_:of in 4394) [ClassicSimilarity], result of:
          0.019723112 = score(doc=4394,freq=24.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.3591007 = fieldWeight in 4394, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=4394)
        0.021686744 = weight(_text_:the in 4394) [ClassicSimilarity], result of:
          0.021686744 = score(doc=4394,freq=28.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.39134735 = fieldWeight in 4394, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=4394)
      0.5 = coord(6/12)
    
    Abstract
    Purpose - To propose a categorization of the different conflation procedures at the two basic approaches, non-linguistic and linguistic techniques, and to justify the application of normalization methods within the framework of linguistic techniques. Design/methodology/approach - Presents a range of term conflation methods, that can be used in information retrieval. The uniterm and multiterm variants can be considered equivalent units for the purposes of automatic indexing. Stemming algorithms, segmentation rules, association measures and clustering techniques are well evaluated non-linguistic methods, and experiments with these techniques show a wide variety of results. Alternatively, the lemmatisation and the use of syntactic pattern-matching, through equivalence relations represented in finite-state transducers (FST), are emerging methods for the recognition and standardization of terms. Findings - The survey attempts to point out the positive and negative effects of the linguistic approach and its potential as a term conflation method. Originality/value - Outlines the importance of FSTs for the normalization of term variants.
    Source
    Journal of documentation. 61(2005) no.4, S.520-547
  17. Rahmstorf, G.: Information retrieval using conceptual representations of phrases (1994) 0.05
    0.04885847 = product of:
      0.09771694 = sum of:
        0.01898392 = weight(_text_:information in 7862) [ClassicSimilarity], result of:
          0.01898392 = score(doc=7862,freq=14.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.3078936 = fieldWeight in 7862, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=7862)
        0.011437457 = weight(_text_:und in 7862) [ClassicSimilarity], result of:
          0.011437457 = score(doc=7862,freq=2.0), product of:
            0.07784514 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.035122856 = queryNorm
            0.14692576 = fieldWeight in 7862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=7862)
        0.016415559 = weight(_text_:for in 7862) [ClassicSimilarity], result of:
          0.016415559 = score(doc=7862,freq=8.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.24892932 = fieldWeight in 7862, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=7862)
        0.01738808 = weight(_text_:the in 7862) [ClassicSimilarity], result of:
          0.01738808 = score(doc=7862,freq=18.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.31377596 = fieldWeight in 7862, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=7862)
        0.016103853 = weight(_text_:of in 7862) [ClassicSimilarity], result of:
          0.016103853 = score(doc=7862,freq=16.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.2932045 = fieldWeight in 7862, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=7862)
        0.01738808 = weight(_text_:the in 7862) [ClassicSimilarity], result of:
          0.01738808 = score(doc=7862,freq=18.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.31377596 = fieldWeight in 7862, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=7862)
      0.5 = coord(6/12)
    
    Abstract
    The information retrieval problem is described starting from an analysis of the concepts 'user's information request' and 'information offerings of texts'. It is shown that natural language phrases are a more adequate medium for expressing information requests and information offerings than character string based query and indexing languages complemented by Boolean oprators. The phrases must be represented as concepts to reach a language invariant level for rule based relevance analysis. The special type of representation called advanced thesaurus is used for the semantic representation of natural language phrases and for relevance processing. The analysis of the retrieval problem leads to a symmetric system structure
    Source
    Information systems and data analysis: prospects - foundations - applications. Proc. of the 17th Annual Conference of the Gesellschaft für Klassifikation, Kaiserslautern, March 3-5, 1993. Ed.: H.-H. Bock et al
    Theme
    Konzeption und Anwendung des Prinzips Thesaurus
  18. Ahmed, F.; Nürnberger, A.: Evaluation of n-gram conflation approaches for Arabic text retrieval (2009) 0.05
    0.04841123 = product of:
      0.09682246 = sum of:
        0.012329709 = product of:
          0.036989126 = sum of:
            0.036989126 = weight(_text_:f in 2941) [ClassicSimilarity], result of:
              0.036989126 = score(doc=2941,freq=2.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.26422277 = fieldWeight in 2941, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2941)
          0.33333334 = coord(1/3)
        0.007175247 = weight(_text_:information in 2941) [ClassicSimilarity], result of:
          0.007175247 = score(doc=2941,freq=2.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.116372846 = fieldWeight in 2941, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2941)
        0.023215106 = weight(_text_:for in 2941) [ClassicSimilarity], result of:
          0.023215106 = score(doc=2941,freq=16.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.35203922 = fieldWeight in 2941, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.046875 = fieldNorm(doc=2941)
        0.020078024 = weight(_text_:the in 2941) [ClassicSimilarity], result of:
          0.020078024 = score(doc=2941,freq=24.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.36231726 = fieldWeight in 2941, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=2941)
        0.013946345 = weight(_text_:of in 2941) [ClassicSimilarity], result of:
          0.013946345 = score(doc=2941,freq=12.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.25392252 = fieldWeight in 2941, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=2941)
        0.020078024 = weight(_text_:the in 2941) [ClassicSimilarity], result of:
          0.020078024 = score(doc=2941,freq=24.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.36231726 = fieldWeight in 2941, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.046875 = fieldNorm(doc=2941)
      0.5 = coord(6/12)
    
    Abstract
    In this paper we present a language-independent approach for conflation that does not depend on predefined rules or prior knowledge of the target language. The proposed unsupervised method is based on an enhancement of the pure n-gram model that can group related words based on various string-similarity measures, while restricting the search to specific locations of the target word by taking into account the order of n-grams. We show that the method is effective to achieve high score similarities for all word-form variations and reduces the ambiguity, i.e., obtains a higher precision and recall, compared to pure n-gram-based approaches for English, Portuguese, and Arabic. The proposed method is especially suited for conflation approaches in Arabic, since Arabic is a highly inflectional language. Therefore, we present in addition an adaptive user interface for Arabic text retrieval called araSearch. araSearch serves as a metasearch interface to existing search engines. The system is able to extend a query using the proposed conflation approach such that additional results for relevant subwords can be found automatically.
    Source
    Journal of the American Society for Information Science and Technology. 60(2009) no.7, S.1448-1465
  19. Shaalan, K.; Raza, H.: NERA: Named Entity Recognition for Arabic (2009) 0.05
    0.048265092 = product of:
      0.096530184 = sum of:
        0.010274758 = product of:
          0.030824272 = sum of:
            0.030824272 = weight(_text_:f in 2953) [ClassicSimilarity], result of:
              0.030824272 = score(doc=2953,freq=2.0), product of:
                0.13999219 = queryWeight, product of:
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.035122856 = queryNorm
                0.22018565 = fieldWeight in 2953, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.985786 = idf(docFreq=2232, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2953)
          0.33333334 = coord(1/3)
        0.005979372 = weight(_text_:information in 2953) [ClassicSimilarity], result of:
          0.005979372 = score(doc=2953,freq=2.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.09697737 = fieldWeight in 2953, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2953)
        0.018096453 = weight(_text_:for in 2953) [ClassicSimilarity], result of:
          0.018096453 = score(doc=2953,freq=14.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.27441877 = fieldWeight in 2953, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2953)
        0.021600515 = weight(_text_:the in 2953) [ClassicSimilarity], result of:
          0.021600515 = score(doc=2953,freq=40.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3897913 = fieldWeight in 2953, product of:
              6.3245554 = tf(freq=40.0), with freq of:
                40.0 = termFreq=40.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2953)
        0.018978573 = weight(_text_:of in 2953) [ClassicSimilarity], result of:
          0.018978573 = score(doc=2953,freq=32.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.34554482 = fieldWeight in 2953, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2953)
        0.021600515 = weight(_text_:the in 2953) [ClassicSimilarity], result of:
          0.021600515 = score(doc=2953,freq=40.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.3897913 = fieldWeight in 2953, product of:
              6.3245554 = tf(freq=40.0), with freq of:
                40.0 = termFreq=40.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2953)
      0.5 = coord(6/12)
    
    Abstract
    Name identification has been worked on quite intensively for the past few years, and has been incorporated into several products revolving around natural language processing tasks. Many researchers have attacked the name identification problem in a variety of languages, but only a few limited research efforts have focused on named entity recognition for Arabic script. This is due to the lack of resources for Arabic named entities and the limited amount of progress made in Arabic natural language processing in general. In this article, we present the results of our attempt at the recognition and extraction of the 10 most important categories of named entities in Arabic script: the person name, location, company, date, time, price, measurement, phone number, ISBN, and file name. We developed the system Named Entity Recognition for Arabic (NERA) using a rule-based approach. The resources created are: a Whitelist representing a dictionary of names, and a grammar, in the form of regular expressions, which are responsible for recognizing the named entities. A filtration mechanism is used that serves two different purposes: (a) revision of the results from a named entity extractor by using metadata, in terms of a Blacklist or rejecter, about ill-formed named entities and (b) disambiguation of identical or overlapping textual matches returned by different name entity extractors to get the correct choice. In NERA, we addressed major challenges posed by NER in the Arabic language arising due to the complexity of the language, peculiarities in the Arabic orthographic system, nonstandardization of the written text, ambiguity, and lack of resources. NERA has been effectively evaluated using our own tagged corpus; it achieved satisfactory results in terms of precision, recall, and F-measure.
    Source
    Journal of the American Society for Information Science and Technology. 60(2009) no.8, S.1652-1663
  20. Riloff, E.: ¬An empirical study of automated dictionary construction for information extraction in three domains (1996) 0.05
    0.048104133 = product of:
      0.09620827 = sum of:
        0.016570523 = weight(_text_:information in 6752) [ClassicSimilarity], result of:
          0.016570523 = score(doc=6752,freq=6.0), product of:
            0.0616574 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.035122856 = queryNorm
            0.2687516 = fieldWeight in 6752, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=6752)
        0.018955056 = weight(_text_:for in 6752) [ClassicSimilarity], result of:
          0.018955056 = score(doc=6752,freq=6.0), product of:
            0.06594466 = queryWeight, product of:
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.035122856 = queryNorm
            0.28743884 = fieldWeight in 6752, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.8775425 = idf(docFreq=18385, maxDocs=44218)
              0.0625 = fieldNorm(doc=6752)
        0.01545607 = weight(_text_:the in 6752) [ClassicSimilarity], result of:
          0.01545607 = score(doc=6752,freq=8.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.27891195 = fieldWeight in 6752, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0625 = fieldNorm(doc=6752)
        0.010735902 = weight(_text_:of in 6752) [ClassicSimilarity], result of:
          0.010735902 = score(doc=6752,freq=4.0), product of:
            0.054923624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.035122856 = queryNorm
            0.19546966 = fieldWeight in 6752, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=6752)
        0.01545607 = weight(_text_:the in 6752) [ClassicSimilarity], result of:
          0.01545607 = score(doc=6752,freq=8.0), product of:
            0.05541559 = queryWeight, product of:
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.035122856 = queryNorm
            0.27891195 = fieldWeight in 6752, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5777643 = idf(docFreq=24812, maxDocs=44218)
              0.0625 = fieldNorm(doc=6752)
        0.019034648 = product of:
          0.038069297 = sum of:
            0.038069297 = weight(_text_:22 in 6752) [ClassicSimilarity], result of:
              0.038069297 = score(doc=6752,freq=2.0), product of:
                0.12299426 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.035122856 = queryNorm
                0.30952093 = fieldWeight in 6752, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6752)
          0.5 = coord(1/2)
      0.5 = coord(6/12)
    
    Abstract
    AutoSlog is a system that addresses the knowledge engineering bottleneck for information extraction. AutoSlog automatically creates domain specific dictionaries for information extraction, given an appropriate training corpus. Describes experiments with AutoSlog in terrorism, joint ventures and microelectronics domains. Compares the performance of AutoSlog across the 3 domains, discusses the lessons learned and presents results from 2 experiments which demonstrate that novice users can generate effective dictionaries using AutoSlog
    Date
    6. 3.1997 16:22:15

Languages

Types

  • a 603
  • el 87
  • m 79
  • s 31
  • x 15
  • p 7
  • b 3
  • d 3
  • n 1
  • pat 1
  • r 1
  • More… Less…

Subjects

Classifications