-
Chen, L.; Fang, H.: ¬An automatic method for ex-tracting innovative ideas based on the Scopus® database (2019)
0.03
0.03380641 = product of:
0.08451602 = sum of:
0.002284684 = product of:
0.004569368 = sum of:
0.004569368 = weight(_text_:e in 5310) [ClassicSimilarity], result of:
0.004569368 = score(doc=5310,freq=2.0), product of:
0.05754566 = queryWeight, product of:
1.43737 = idf(docFreq=28552, maxDocs=44218)
0.040035386 = queryNorm
0.07940422 = fieldWeight in 5310, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
1.43737 = idf(docFreq=28552, maxDocs=44218)
0.0390625 = fieldNorm(doc=5310)
0.5 = coord(1/2)
0.082231335 = weight(_text_:n in 5310) [ClassicSimilarity], result of:
0.082231335 = score(doc=5310,freq=8.0), product of:
0.17261884 = queryWeight, product of:
4.3116565 = idf(docFreq=1611, maxDocs=44218)
0.040035386 = queryNorm
0.47637522 = fieldWeight in 5310, product of:
2.828427 = tf(freq=8.0), with freq of:
8.0 = termFreq=8.0
4.3116565 = idf(docFreq=1611, maxDocs=44218)
0.0390625 = fieldNorm(doc=5310)
0.4 = coord(2/5)
- Abstract
- The novelty of knowledge claims in a research paper can be considered an evaluation criterion for papers to supplement citations. To provide a foundation for research evaluation from the perspective of innovativeness, we propose an automatic approach for extracting innovative ideas from the abstracts of technology and engineering papers. The approach extracts N-grams as candidates based on part-of-speech tagging and determines whether they are novel by checking the Scopus® database to determine whether they had ever been presented previously. Moreover, we discussed the distributions of innovative ideas in different abstract structures. To improve the performance by excluding noisy N-grams, a list of stopwords and a list of research description characteristics were developed. We selected abstracts of articles published from 2011 to 2017 with the topic of semantic analysis as the experimental texts. Excluding noisy N-grams, considering the distribution of innovative ideas in abstracts, and suitably combining N-grams can effectively improve the performance of automatic innovative idea extraction. Unlike co-word and co-citation analysis, innovative-idea extraction aims to identify the differences in a paper from all previously published papers.
- Language
- e
-
He, Q.: Knowledge discovery through co-word analysis (1999)
0.00
0.001279423 = product of:
0.006397115 = sum of:
0.006397115 = product of:
0.01279423 = sum of:
0.01279423 = weight(_text_:e in 6082) [ClassicSimilarity], result of:
0.01279423 = score(doc=6082,freq=2.0), product of:
0.05754566 = queryWeight, product of:
1.43737 = idf(docFreq=28552, maxDocs=44218)
0.040035386 = queryNorm
0.2223318 = fieldWeight in 6082, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
1.43737 = idf(docFreq=28552, maxDocs=44218)
0.109375 = fieldNorm(doc=6082)
0.5 = coord(1/2)
0.2 = coord(1/5)
- Language
- e
-
Ahonen, H.: Knowledge discovery in documents by extracting frequent word sequences (1999)
0.00
0.001279423 = product of:
0.006397115 = sum of:
0.006397115 = product of:
0.01279423 = sum of:
0.01279423 = weight(_text_:e in 6088) [ClassicSimilarity], result of:
0.01279423 = score(doc=6088,freq=2.0), product of:
0.05754566 = queryWeight, product of:
1.43737 = idf(docFreq=28552, maxDocs=44218)
0.040035386 = queryNorm
0.2223318 = fieldWeight in 6088, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
1.43737 = idf(docFreq=28552, maxDocs=44218)
0.109375 = fieldNorm(doc=6088)
0.5 = coord(1/2)
0.2 = coord(1/5)
- Language
- e
-
He, Q.: ¬A study of the strength indexes in co-word analysis (2000)
0.00
9.4972545E-4 = product of:
0.004748627 = sum of:
0.004748627 = product of:
0.009497254 = sum of:
0.009497254 = weight(_text_:e in 111) [ClassicSimilarity], result of:
0.009497254 = score(doc=111,freq=6.0), product of:
0.05754566 = queryWeight, product of:
1.43737 = idf(docFreq=28552, maxDocs=44218)
0.040035386 = queryNorm
0.16503859 = fieldWeight in 111, product of:
2.4494898 = tf(freq=6.0), with freq of:
6.0 = termFreq=6.0
1.43737 = idf(docFreq=28552, maxDocs=44218)
0.046875 = fieldNorm(doc=111)
0.5 = coord(1/2)
0.2 = coord(1/5)
- Abstract
- Co-word analysis is a technique for detecting the knowledge structure of scientific literature and mapping the dynamics in a research field. It is used to count the co-occurrences of term pairs, compute the strength between term pairs, and map the research field by inserting terms and their linkages into a graphical structure according to the strength values. In previous co-word studies, there are two indexes used to measure the strength between term pairs in order to identify the major areas in a research field - the inclusion index (I) and the equivalence index (E). This study will conduct two co-word analysis experiments using the two indexes, respectively, and compare the results from the two experiments. The results show, due to the difference in their computation, index I is more likely to identify general subject areas in a research field while index E is more likely to identify subject areas at more specific levels
- Language
- e
-
Radev, D.R.; Joseph, M.T.; Gibson, B.; Muthukrishnan, P.: ¬A bibliometric and network analysis of the field of computational linguistics (2016)
0.00
6.397115E-4 = product of:
0.0031985575 = sum of:
0.0031985575 = product of:
0.006397115 = sum of:
0.006397115 = weight(_text_:e in 2764) [ClassicSimilarity], result of:
0.006397115 = score(doc=2764,freq=2.0), product of:
0.05754566 = queryWeight, product of:
1.43737 = idf(docFreq=28552, maxDocs=44218)
0.040035386 = queryNorm
0.1111659 = fieldWeight in 2764, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
1.43737 = idf(docFreq=28552, maxDocs=44218)
0.0546875 = fieldNorm(doc=2764)
0.5 = coord(1/2)
0.2 = coord(1/5)
- Language
- e
-
Moohebat, M.; Raj, R.G.; Kareem, S.B.A.; Thorleuchter, D.: Identifying ISI-indexed articles by their lexical usage : a text analysis approach (2015)
0.00
5.4832414E-4 = product of:
0.0027416206 = sum of:
0.0027416206 = product of:
0.0054832413 = sum of:
0.0054832413 = weight(_text_:e in 1664) [ClassicSimilarity], result of:
0.0054832413 = score(doc=1664,freq=2.0), product of:
0.05754566 = queryWeight, product of:
1.43737 = idf(docFreq=28552, maxDocs=44218)
0.040035386 = queryNorm
0.09528506 = fieldWeight in 1664, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
1.43737 = idf(docFreq=28552, maxDocs=44218)
0.046875 = fieldNorm(doc=1664)
0.5 = coord(1/2)
0.2 = coord(1/5)
- Language
- e
-
Levin, M.; Krawczyk, S.; Bethard, S.; Jurafsky, D.: Citation-based bootstrapping for large-scale author disambiguation (2012)
0.00
4.569368E-4 = product of:
0.002284684 = sum of:
0.002284684 = product of:
0.004569368 = sum of:
0.004569368 = weight(_text_:e in 246) [ClassicSimilarity], result of:
0.004569368 = score(doc=246,freq=2.0), product of:
0.05754566 = queryWeight, product of:
1.43737 = idf(docFreq=28552, maxDocs=44218)
0.040035386 = queryNorm
0.07940422 = fieldWeight in 246, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
1.43737 = idf(docFreq=28552, maxDocs=44218)
0.0390625 = fieldNorm(doc=246)
0.5 = coord(1/2)
0.2 = coord(1/5)
- Language
- e