Search (4 results, page 1 of 1)

  • × theme_ss:"Citation indexing"
  • × year_i:[2020 TO 2030}
  1. Tay, A.: ¬The next generation discovery citation indexes : a review of the landscape in 2020 (2020) 0.01
    0.010864421 = product of:
      0.021728842 = sum of:
        0.021728842 = product of:
          0.043457683 = sum of:
            0.043457683 = weight(_text_:22 in 40) [ClassicSimilarity], result of:
              0.043457683 = score(doc=40,freq=2.0), product of:
                0.16046064 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045821942 = queryNorm
                0.2708308 = fieldWeight in 40, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=40)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    17.11.2020 12:22:59
  2. Xie, J.; Lu, H.; Kang, L.; Cheng, Y.: Citing criteria and its effects on researcher's intention to cite : a mixed-method study (2022) 0.00
    0.004687378 = product of:
      0.009374756 = sum of:
        0.009374756 = product of:
          0.018749513 = sum of:
            0.018749513 = weight(_text_:h in 634) [ClassicSimilarity], result of:
              0.018749513 = score(doc=634,freq=2.0), product of:
                0.113842286 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045821942 = queryNorm
                0.16469726 = fieldWeight in 634, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.046875 = fieldNorm(doc=634)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  3. Cui, Y.; Wang, Y.; Liu, X.; Wang, X.; Zhang, X.: Multidimensional scholarly citations : characterizing and understanding scholars' citation behaviors (2023) 0.00
    0.0039061487 = product of:
      0.0078122974 = sum of:
        0.0078122974 = product of:
          0.015624595 = sum of:
            0.015624595 = weight(_text_:h in 847) [ClassicSimilarity], result of:
              0.015624595 = score(doc=847,freq=2.0), product of:
                0.113842286 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045821942 = queryNorm
                0.13724773 = fieldWeight in 847, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=847)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This study investigates scholars' citation behaviors from a fine-grained perspective. Specifically, each scholarly citation is considered multidimensional rather than logically unidimensional (i.e., present or absent). Thirty million articles from PubMed were accessed for use in empirical research, in which a total of 15 interpretable features of scholarly citations were constructed and grouped into three main categories. Each category corresponds to one aspect of the reasons and motivations behind scholars' citation decision-making during academic writing. Using about 500,000 pairs of actual and randomly generated scholarly citations, a series of Random Forest-based classification experiments were conducted to quantitatively evaluate the correlation between each constructed citation feature and citation decisions made by scholars. Our experimental results indicate that citation proximity is the category most relevant to scholars' citation decision-making, followed by citation authority and citation inertia. However, big-name scholars whose h-indexes rank among the top 1% exhibit a unique pattern of citation behaviors-their citation decision-making correlates most closely with citation inertia, with the correlation nearly three times as strong as that of their ordinary counterparts. Hopefully, the empirical findings presented in this paper can bring us closer to characterizing and understanding the complex process of generating scholarly citations in academia.
  4. Safder, I.; Ali, M.; Aljohani, N.R.; Nawaz, R.; Hassan, S.-U.: Neural machine translation for in-text citation classification (2023) 0.00
    0.0039061487 = product of:
      0.0078122974 = sum of:
        0.0078122974 = product of:
          0.015624595 = sum of:
            0.015624595 = weight(_text_:h in 1053) [ClassicSimilarity], result of:
              0.015624595 = score(doc=1053,freq=2.0), product of:
                0.113842286 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045821942 = queryNorm
                0.13724773 = fieldWeight in 1053, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1053)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The quality of scientific publications can be measured by quantitative indices such as the h-index, Source Normalized Impact per Paper, or g-index. However, these measures lack to explain the function or reasons for citations and the context of citations from citing publication to cited publication. We argue that citation context may be considered while calculating the impact of research work. However, mining citation context from unstructured full-text publications is a challenging task. In this paper, we compiled a data set comprising 9,518 citations context. We developed a deep learning-based architecture for citation context classification. Unlike feature-based state-of-the-art models, our proposed focal-loss and class-weight-aware BiLSTM model with pretrained GloVe embedding vectors use citation context as input to outperform them in multiclass citation context classification tasks. Our model improves on the baseline state-of-the-art by achieving an F1 score of 0.80 with an accuracy of 0.81 for citation context classification. Moreover, we delve into the effects of using different word embeddings on the performance of the classification model and draw a comparison between fastText, GloVe, and spaCy pretrained word embeddings.