Search (7 results, page 1 of 1)

  • × theme_ss:"Computerlinguistik"
  • × theme_ss:"Informetrie"
  • × type_ss:"a"
  1. Chen, L.; Fang, H.: ¬An automatic method for ex-tracting innovative ideas based on the Scopus® database (2019) 0.03
    0.03380641 = product of:
      0.08451602 = sum of:
        0.002284684 = product of:
          0.004569368 = sum of:
            0.004569368 = weight(_text_:e in 5310) [ClassicSimilarity], result of:
              0.004569368 = score(doc=5310,freq=2.0), product of:
                0.05754566 = queryWeight, product of:
                  1.43737 = idf(docFreq=28552, maxDocs=44218)
                  0.040035386 = queryNorm
                0.07940422 = fieldWeight in 5310, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.43737 = idf(docFreq=28552, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5310)
          0.5 = coord(1/2)
        0.082231335 = weight(_text_:n in 5310) [ClassicSimilarity], result of:
          0.082231335 = score(doc=5310,freq=8.0), product of:
            0.17261884 = queryWeight, product of:
              4.3116565 = idf(docFreq=1611, maxDocs=44218)
              0.040035386 = queryNorm
            0.47637522 = fieldWeight in 5310, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              4.3116565 = idf(docFreq=1611, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5310)
      0.4 = coord(2/5)
    
    Abstract
    The novelty of knowledge claims in a research paper can be considered an evaluation criterion for papers to supplement citations. To provide a foundation for research evaluation from the perspective of innovativeness, we propose an automatic approach for extracting innovative ideas from the abstracts of technology and engineering papers. The approach extracts N-grams as candidates based on part-of-speech tagging and determines whether they are novel by checking the Scopus® database to determine whether they had ever been presented previously. Moreover, we discussed the distributions of innovative ideas in different abstract structures. To improve the performance by excluding noisy N-grams, a list of stopwords and a list of research description characteristics were developed. We selected abstracts of articles published from 2011 to 2017 with the topic of semantic analysis as the experimental texts. Excluding noisy N-grams, considering the distribution of innovative ideas in abstracts, and suitably combining N-grams can effectively improve the performance of automatic innovative idea extraction. Unlike co-word and co-citation analysis, innovative-idea extraction aims to identify the differences in a paper from all previously published papers.
    Language
    e
  2. He, Q.: Knowledge discovery through co-word analysis (1999) 0.00
    0.001279423 = product of:
      0.006397115 = sum of:
        0.006397115 = product of:
          0.01279423 = sum of:
            0.01279423 = weight(_text_:e in 6082) [ClassicSimilarity], result of:
              0.01279423 = score(doc=6082,freq=2.0), product of:
                0.05754566 = queryWeight, product of:
                  1.43737 = idf(docFreq=28552, maxDocs=44218)
                  0.040035386 = queryNorm
                0.2223318 = fieldWeight in 6082, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.43737 = idf(docFreq=28552, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6082)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Language
    e
  3. Ahonen, H.: Knowledge discovery in documents by extracting frequent word sequences (1999) 0.00
    0.001279423 = product of:
      0.006397115 = sum of:
        0.006397115 = product of:
          0.01279423 = sum of:
            0.01279423 = weight(_text_:e in 6088) [ClassicSimilarity], result of:
              0.01279423 = score(doc=6088,freq=2.0), product of:
                0.05754566 = queryWeight, product of:
                  1.43737 = idf(docFreq=28552, maxDocs=44218)
                  0.040035386 = queryNorm
                0.2223318 = fieldWeight in 6088, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.43737 = idf(docFreq=28552, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6088)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Language
    e
  4. He, Q.: ¬A study of the strength indexes in co-word analysis (2000) 0.00
    9.4972545E-4 = product of:
      0.004748627 = sum of:
        0.004748627 = product of:
          0.009497254 = sum of:
            0.009497254 = weight(_text_:e in 111) [ClassicSimilarity], result of:
              0.009497254 = score(doc=111,freq=6.0), product of:
                0.05754566 = queryWeight, product of:
                  1.43737 = idf(docFreq=28552, maxDocs=44218)
                  0.040035386 = queryNorm
                0.16503859 = fieldWeight in 111, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.43737 = idf(docFreq=28552, maxDocs=44218)
                  0.046875 = fieldNorm(doc=111)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Abstract
    Co-word analysis is a technique for detecting the knowledge structure of scientific literature and mapping the dynamics in a research field. It is used to count the co-occurrences of term pairs, compute the strength between term pairs, and map the research field by inserting terms and their linkages into a graphical structure according to the strength values. In previous co-word studies, there are two indexes used to measure the strength between term pairs in order to identify the major areas in a research field - the inclusion index (I) and the equivalence index (E). This study will conduct two co-word analysis experiments using the two indexes, respectively, and compare the results from the two experiments. The results show, due to the difference in their computation, index I is more likely to identify general subject areas in a research field while index E is more likely to identify subject areas at more specific levels
    Language
    e
  5. Radev, D.R.; Joseph, M.T.; Gibson, B.; Muthukrishnan, P.: ¬A bibliometric and network analysis of the field of computational linguistics (2016) 0.00
    6.397115E-4 = product of:
      0.0031985575 = sum of:
        0.0031985575 = product of:
          0.006397115 = sum of:
            0.006397115 = weight(_text_:e in 2764) [ClassicSimilarity], result of:
              0.006397115 = score(doc=2764,freq=2.0), product of:
                0.05754566 = queryWeight, product of:
                  1.43737 = idf(docFreq=28552, maxDocs=44218)
                  0.040035386 = queryNorm
                0.1111659 = fieldWeight in 2764, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.43737 = idf(docFreq=28552, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2764)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Language
    e
  6. Moohebat, M.; Raj, R.G.; Kareem, S.B.A.; Thorleuchter, D.: Identifying ISI-indexed articles by their lexical usage : a text analysis approach (2015) 0.00
    5.4832414E-4 = product of:
      0.0027416206 = sum of:
        0.0027416206 = product of:
          0.0054832413 = sum of:
            0.0054832413 = weight(_text_:e in 1664) [ClassicSimilarity], result of:
              0.0054832413 = score(doc=1664,freq=2.0), product of:
                0.05754566 = queryWeight, product of:
                  1.43737 = idf(docFreq=28552, maxDocs=44218)
                  0.040035386 = queryNorm
                0.09528506 = fieldWeight in 1664, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.43737 = idf(docFreq=28552, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1664)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Language
    e
  7. Levin, M.; Krawczyk, S.; Bethard, S.; Jurafsky, D.: Citation-based bootstrapping for large-scale author disambiguation (2012) 0.00
    4.569368E-4 = product of:
      0.002284684 = sum of:
        0.002284684 = product of:
          0.004569368 = sum of:
            0.004569368 = weight(_text_:e in 246) [ClassicSimilarity], result of:
              0.004569368 = score(doc=246,freq=2.0), product of:
                0.05754566 = queryWeight, product of:
                  1.43737 = idf(docFreq=28552, maxDocs=44218)
                  0.040035386 = queryNorm
                0.07940422 = fieldWeight in 246, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.43737 = idf(docFreq=28552, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=246)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Language
    e