Search (64 results, page 1 of 4)

  • × theme_ss:"Computerlinguistik"
  1. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.07
    0.07155078 = sum of:
      0.05334761 = product of:
        0.21339044 = sum of:
          0.21339044 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
            0.21339044 = score(doc=562,freq=2.0), product of:
              0.37968612 = queryWeight, product of:
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.04478481 = queryNorm
              0.56201804 = fieldWeight in 562, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.046875 = fieldNorm(doc=562)
        0.25 = coord(1/4)
      0.018203171 = product of:
        0.036406342 = sum of:
          0.036406342 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
            0.036406342 = score(doc=562,freq=2.0), product of:
              0.15682878 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04478481 = queryNorm
              0.23214069 = fieldWeight in 562, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=562)
        0.5 = coord(1/2)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  2. Doszkocs, T.E.; Zamora, A.: Dictionary services and spelling aids for Web searching (2004) 0.06
    0.056691445 = product of:
      0.11338289 = sum of:
        0.11338289 = sum of:
          0.070477605 = weight(_text_:z in 2541) [ClassicSimilarity], result of:
            0.070477605 = score(doc=2541,freq=2.0), product of:
              0.23903055 = queryWeight, product of:
                5.337313 = idf(docFreq=577, maxDocs=44218)
                0.04478481 = queryNorm
              0.29484767 = fieldWeight in 2541, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.337313 = idf(docFreq=577, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2541)
          0.042905286 = weight(_text_:22 in 2541) [ClassicSimilarity], result of:
            0.042905286 = score(doc=2541,freq=4.0), product of:
              0.15682878 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04478481 = queryNorm
              0.27358043 = fieldWeight in 2541, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2541)
      0.5 = coord(1/2)
    
    Abstract
    The Specialized Information Services Division (SIS) of the National Library of Medicine (NLM) provides Web access to more than a dozen scientific databases on toxicology and the environment on TOXNET . Search queries on TOXNET often include misspelled or variant English words, medical and scientific jargon and chemical names. Following the example of search engines like Google and ClinicalTrials.gov, we set out to develop a spelling "suggestion" system for increased recall and precision in TOXNET searching. This paper describes development of dictionary technology that can be used in a variety of applications such as orthographic verification, writing aid, natural language processing, and information storage and retrieval. The design of the technology allows building complex applications using the components developed in the earlier phases of the work in a modular fashion without extensive rewriting of computer code. Since many of the potential applications envisioned for this work have on-line or web-based interfaces, the dictionaries and other computer components must have fast response, and must be adaptable to open-ended database vocabularies, including chemical nomenclature. The dictionary vocabulary for this work was derived from SIS and other databases and specialized resources, such as NLM's Unified Medical Language Systems (UMLS) . The resulting technology, A-Z Dictionary (AZdict), has three major constituents: 1) the vocabulary list, 2) the word attributes that define part of speech and morphological relationships between words in the list, and 3) a set of programs that implements the retrieval of words and their attributes, and determines similarity between words (ChemSpell). These three components can be used in various applications such as spelling verification, spelling aid, part-of-speech tagging, paraphrasing, and many other natural language processing functions.
    Date
    14. 8.2004 17:22:56
    Source
    Online. 28(2004) no.3, S.22-29
  3. Sprachtechnologie : ein Überblick (2012) 0.04
    0.035238802 = product of:
      0.070477605 = sum of:
        0.070477605 = product of:
          0.14095521 = sum of:
            0.14095521 = weight(_text_:z in 1750) [ClassicSimilarity], result of:
              0.14095521 = score(doc=1750,freq=8.0), product of:
                0.23903055 = queryWeight, product of:
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.04478481 = queryNorm
                0.58969533 = fieldWeight in 1750, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1750)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Seit mehr als einem halben Jahrhundert existieren ernsthafte und ernst zu nehmende Versuche, menschliche Sprache maschinell zu verarbeiten. Maschinelle Übersetzung oder "natürliche" Dialoge mit Computern gehören zu den ersten Ideen, die den Bereich der späteren Computerlinguistik oder Sprachtechnologie abgesteckt und deren Vorhaben geleitet haben. Heute ist dieser auch maschinelle Sprachverarbeitung (natural language processing, NLP) genannte Bereich stark ausdiversifiziert: Durch die rapide Entwicklung der Informatik ist vieles vorher Unvorstellbare Realität (z. B. automatische Telefonauskunft), einiges früher Unmögliche immerhin möglich geworden (z. B. Handhelds mit Sprachein- und -ausgabe als digitale persönliche (Informations-)Assistenten). Es gibt verschiedene Anwendungen der Computerlinguistik, von denen einige den Sprung in die kommerzielle Nutzung geschafft haben (z. B. Diktiersysteme, Textklassifikation, maschinelle Übersetzung). Immer noch wird an natürlichsprachlichen Systemen (natural language systems, NLS) verschiedenster Funktionalität (z. B. zur Beantwortung beliebiger Fragen oder zur Generierung komplexer Texte) intensiv geforscht, auch wenn die hoch gesteckten Ziele von einst längst nicht erreicht sind (und deshalb entsprechend "heruntergefahren" wurden). Wo die maschinelle Sprachverarbeitung heute steht, ist allerdings angesichts der vielfältigen Aktivitäten in der Computerlinguistik und Sprachtechnologie weder offensichtlich noch leicht in Erfahrung zu bringen (für Studierende des Fachs und erst recht für Laien). Ein Ziel dieses Buches ist, es, die aktuelle Literaturlage in dieser Hinsicht zu verbessern, indem spezifisch systembezogene Aspekte der Computerlinguistik als Überblick über die Sprachtechnologie zusammengetragen werden.
  4. Xianghao, G.; Yixin, Z.; Li, Y.: ¬A new method of news test understanding and abstracting based on speech acts theory (1998) 0.03
    0.028191041 = product of:
      0.056382082 = sum of:
        0.056382082 = product of:
          0.112764165 = sum of:
            0.112764165 = weight(_text_:z in 3532) [ClassicSimilarity], result of:
              0.112764165 = score(doc=3532,freq=2.0), product of:
                0.23903055 = queryWeight, product of:
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.04478481 = queryNorm
                0.47175628 = fieldWeight in 3532, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3532)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  5. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.03
    0.026673805 = product of:
      0.05334761 = sum of:
        0.05334761 = product of:
          0.21339044 = sum of:
            0.21339044 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
              0.21339044 = score(doc=862,freq=2.0), product of:
                0.37968612 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.04478481 = queryNorm
                0.56201804 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=862)
          0.25 = coord(1/4)
      0.5 = coord(1/2)
    
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
  6. Savoy, J.: Text representation strategies : an example with the State of the union addresses (2016) 0.02
    0.024917595 = product of:
      0.04983519 = sum of:
        0.04983519 = product of:
          0.09967038 = sum of:
            0.09967038 = weight(_text_:z in 3042) [ClassicSimilarity], result of:
              0.09967038 = score(doc=3042,freq=4.0), product of:
                0.23903055 = queryWeight, product of:
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.04478481 = queryNorm
                0.41697758 = fieldWeight in 3042, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3042)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Based on State of the Union addresses from 1790 to 2014 (225 speeches delivered by 42 presidents), this paper describes and evaluates different text representation strategies. To determine the most important words of a given text, the term frequencies (tf) or the tf?idf weighting scheme can be applied. Recently, latent Dirichlet allocation (LDA) has been proposed to define the topics included in a corpus. As another strategy, this study proposes to apply a vocabulary specificity measure (Z?score) to determine the most significantly overused word-types or short sequences of them. Our experiments show that the simple term frequency measure is not able to discriminate between specific terms associated with a document or a set of texts. Using the tf idf or LDA approach, the selection requires some arbitrary decisions. Based on the term-specific measure (Z?score), the term selection has a clear theoretical basis. Moreover, the most significant sentences for each presidency can be determined. As another facet, we can visualize the dynamic evolution of usage of some terms associated with their specificity measures. Finally, this technique can be employed to define the most important lexical leaders introducing terms overused by the k following presidencies.
  7. Granitzer, M.: Statistische Verfahren der Textanalyse (2006) 0.02
    0.024667162 = product of:
      0.049334325 = sum of:
        0.049334325 = product of:
          0.09866865 = sum of:
            0.09866865 = weight(_text_:z in 5809) [ClassicSimilarity], result of:
              0.09866865 = score(doc=5809,freq=2.0), product of:
                0.23903055 = queryWeight, product of:
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.04478481 = queryNorm
                0.41278675 = fieldWeight in 5809, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5809)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Der vorliegende Artikel bietet einen Überblick über statistische Verfahren der Textanalyse im Kontext des Semantic Webs. Als Einleitung erfolgt die Diskussion von Methoden und gängigen Techniken zur Vorverarbeitung von Texten wie z. B. Stemming oder Part-of-Speech Tagging. Die so eingeführten Repräsentationsformen dienen als Basis für statistische Merkmalsanalysen sowie für weiterführende Techniken wie Information Extraction und maschinelle Lernverfahren. Die Darstellung dieser speziellen Techniken erfolgt im Überblick, wobei auf die wichtigsten Aspekte in Bezug auf das Semantic Web detailliert eingegangen wird. Die Anwendung der vorgestellten Techniken zur Erstellung und Wartung von Ontologien sowie der Verweis auf weiterführende Literatur bilden den Abschluss dieses Artikels.
  8. Warner, A.J.: Natural language processing (1987) 0.02
    0.024270894 = product of:
      0.048541788 = sum of:
        0.048541788 = product of:
          0.097083576 = sum of:
            0.097083576 = weight(_text_:22 in 337) [ClassicSimilarity], result of:
              0.097083576 = score(doc=337,freq=2.0), product of:
                0.15682878 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04478481 = queryNorm
                0.61904186 = fieldWeight in 337, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=337)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Annual review of information science and technology. 22(1987), S.79-108
  9. McMahon, J.G.; Smith, F.J.: Improved statistical language model performance with automatic generated word hierarchies (1996) 0.02
    0.021237032 = product of:
      0.042474065 = sum of:
        0.042474065 = product of:
          0.08494813 = sum of:
            0.08494813 = weight(_text_:22 in 3164) [ClassicSimilarity], result of:
              0.08494813 = score(doc=3164,freq=2.0), product of:
                0.15682878 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04478481 = queryNorm
                0.5416616 = fieldWeight in 3164, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3164)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Computational linguistics. 22(1996) no.2, S.217-248
  10. Ruge, G.: ¬A spreading activation network for automatic generation of thesaurus relationships (1991) 0.02
    0.021237032 = product of:
      0.042474065 = sum of:
        0.042474065 = product of:
          0.08494813 = sum of:
            0.08494813 = weight(_text_:22 in 4506) [ClassicSimilarity], result of:
              0.08494813 = score(doc=4506,freq=2.0), product of:
                0.15682878 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04478481 = queryNorm
                0.5416616 = fieldWeight in 4506, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4506)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    8.10.2000 11:52:22
  11. Somers, H.: Example-based machine translation : Review article (1999) 0.02
    0.021237032 = product of:
      0.042474065 = sum of:
        0.042474065 = product of:
          0.08494813 = sum of:
            0.08494813 = weight(_text_:22 in 6672) [ClassicSimilarity], result of:
              0.08494813 = score(doc=6672,freq=2.0), product of:
                0.15682878 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04478481 = queryNorm
                0.5416616 = fieldWeight in 6672, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6672)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    31. 7.1996 9:22:19
  12. New tools for human translators (1997) 0.02
    0.021237032 = product of:
      0.042474065 = sum of:
        0.042474065 = product of:
          0.08494813 = sum of:
            0.08494813 = weight(_text_:22 in 1179) [ClassicSimilarity], result of:
              0.08494813 = score(doc=1179,freq=2.0), product of:
                0.15682878 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04478481 = queryNorm
                0.5416616 = fieldWeight in 1179, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1179)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    31. 7.1996 9:22:19
  13. Baayen, R.H.; Lieber, H.: Word frequency distributions and lexical semantics (1997) 0.02
    0.021237032 = product of:
      0.042474065 = sum of:
        0.042474065 = product of:
          0.08494813 = sum of:
            0.08494813 = weight(_text_:22 in 3117) [ClassicSimilarity], result of:
              0.08494813 = score(doc=3117,freq=2.0), product of:
                0.15682878 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04478481 = queryNorm
                0.5416616 = fieldWeight in 3117, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3117)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    28. 2.1999 10:48:22
  14. ¬Der Student aus dem Computer (2023) 0.02
    0.021237032 = product of:
      0.042474065 = sum of:
        0.042474065 = product of:
          0.08494813 = sum of:
            0.08494813 = weight(_text_:22 in 1079) [ClassicSimilarity], result of:
              0.08494813 = score(doc=1079,freq=2.0), product of:
                0.15682878 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04478481 = queryNorm
                0.5416616 = fieldWeight in 1079, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1079)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    27. 1.2023 16:22:55
  15. Santana Suárez, O.; Carreras Riudavets, F.J.; Hernández Figueroa, Z.; González Cabrera, A.C.: Integration of an XML electronic dictionary with linguistic tools for natural language processing (2007) 0.02
    0.021143282 = product of:
      0.042286564 = sum of:
        0.042286564 = product of:
          0.08457313 = sum of:
            0.08457313 = weight(_text_:z in 921) [ClassicSimilarity], result of:
              0.08457313 = score(doc=921,freq=2.0), product of:
                0.23903055 = queryWeight, product of:
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.04478481 = queryNorm
                0.35381722 = fieldWeight in 921, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.046875 = fieldNorm(doc=921)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  16. Endres-Niggemeyer, B.: Thinkie: Lautes Denken mit Spracherkennung (mobil) (2013) 0.02
    0.021143282 = product of:
      0.042286564 = sum of:
        0.042286564 = product of:
          0.08457313 = sum of:
            0.08457313 = weight(_text_:z in 1145) [ClassicSimilarity], result of:
              0.08457313 = score(doc=1145,freq=2.0), product of:
                0.23903055 = queryWeight, product of:
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.04478481 = queryNorm
                0.35381722 = fieldWeight in 1145, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1145)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Lautes Denken ist eine bewährte Methode zum Erforschen kognitiver Prozesse. Es wird in vielen Disziplinen benutzt, z. B. um aufzudecken, welche Erfahrungen Benutzer bei der Interaktion mit Computerschnittstellen machen. Nach einer kurzen Erklärung des Lauten Denkens wird die App Thinkie vorgestellt. Thinkie ist eine mobile Lösung für das Laute Denken auf iPhone und iPad. Die Testperson nimmt auf dem iPhone den Ton auf. Die Spracherkennungssoftware Siri (http://www.apple.com/de/ios/siri/) transkribiert ihn. Parallel wird auf dem iPad oder einem anderen Gerät gefilmt. Auf dem iPad kann man - mit Video im Blick - das Transkript aufarbeiten und interpretieren. Die Textdateien transportiert Thinkie über eine Cloud-Kollektion, die Filme werden mit iTunes übertragen. Thinkie ist noch nicht tauglich für den praktischen Gebrauch. Noch sind die Sequenzen zu kurz, die Siri verarbeiten kann. Das wird sich ändern.
  17. AL-Smadi, M.; Jaradat, Z.; AL-Ayyoub, M.; Jararweh, Y.: Paraphrase identification and semantic text similarity analysis in Arabic news tweets using lexical, syntactic, and semantic features (2017) 0.02
    0.021143282 = product of:
      0.042286564 = sum of:
        0.042286564 = product of:
          0.08457313 = sum of:
            0.08457313 = weight(_text_:z in 5095) [ClassicSimilarity], result of:
              0.08457313 = score(doc=5095,freq=2.0), product of:
                0.23903055 = queryWeight, product of:
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.04478481 = queryNorm
                0.35381722 = fieldWeight in 5095, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5095)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  18. Lund, B.D.; Wang, T.; Mannuru, N.R.; Nie, B.; Shimray, S.; Wang, Z.: ChatGPT and a new academic reality : artificial Intelligence-written research papers and the ethics of the large language models in scholarly publishing (2023) 0.02
    0.021143282 = product of:
      0.042286564 = sum of:
        0.042286564 = product of:
          0.08457313 = sum of:
            0.08457313 = weight(_text_:z in 943) [ClassicSimilarity], result of:
              0.08457313 = score(doc=943,freq=2.0), product of:
                0.23903055 = queryWeight, product of:
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.04478481 = queryNorm
                0.35381722 = fieldWeight in 943, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.046875 = fieldNorm(doc=943)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  19. Meng, K.; Ba, Z.; Ma, Y.; Li, G.: ¬A network coupling approach to detecting hierarchical linkages between science and technology (2024) 0.02
    0.021143282 = product of:
      0.042286564 = sum of:
        0.042286564 = product of:
          0.08457313 = sum of:
            0.08457313 = weight(_text_:z in 1205) [ClassicSimilarity], result of:
              0.08457313 = score(doc=1205,freq=2.0), product of:
                0.23903055 = queryWeight, product of:
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.04478481 = queryNorm
                0.35381722 = fieldWeight in 1205, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.337313 = idf(docFreq=577, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1205)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  20. Byrne, C.C.; McCracken, S.A.: ¬An adaptive thesaurus employing semantic distance, relational inheritance and nominal compound interpretation for linguistic support of information retrieval (1999) 0.02
    0.018203171 = product of:
      0.036406342 = sum of:
        0.036406342 = product of:
          0.072812684 = sum of:
            0.072812684 = weight(_text_:22 in 4483) [ClassicSimilarity], result of:
              0.072812684 = score(doc=4483,freq=2.0), product of:
                0.15682878 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04478481 = queryNorm
                0.46428138 = fieldWeight in 4483, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4483)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    15. 3.2000 10:22:37

Years

Languages

  • e 44
  • d 19
  • chi 1
  • More… Less…

Types

  • a 51
  • el 6
  • m 6
  • s 3
  • p 2
  • x 2
  • d 1
  • More… Less…