Search (67 results, page 2 of 4)

  • × theme_ss:"Informetrie"
  • × year_i:[2020 TO 2030}
  1. Chen, L.; Ding, J.; Larivière, V.: Measuring the citation context of national self-references : how a web journal club is used (2022) 0.00
    0.0022374375 = product of:
      0.004474875 = sum of:
        0.004474875 = product of:
          0.00894975 = sum of:
            0.00894975 = weight(_text_:a in 545) [ClassicSimilarity], result of:
              0.00894975 = score(doc=545,freq=14.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.1685276 = fieldWeight in 545, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=545)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The emphasis on research evaluation has brought scrutiny to the role of self-citations in the scholarly communication process. While author self-citations have been studied at length, little is known on national-level self-references (SRs). This paper analyses the citation context of national SRs, using the full-text of 184,859 papers published in PLOS journals. It investigates the differences between national SRs and nonself-references (NSRs) in terms of their in-text mention, presence in enumerations, and location features. For all countries, national SRs exhibit a higher level of engagement than NSRs. NSRs are more often found in enumerative citances than SRs, which suggests that researchers pay more attention to domestic than foreign studies. There are more mentions of national research in the methods section, which provides evidence that methodologies developed in a nation are more likely to be used by other researchers from the same nation. Publications from the United States are cited at a higher rate in each of the sections, indicating that the country still maintains a dominant position in science. On the whole, this paper contributes to a better understanding of the role of national SRs in the scholarly communication system, and how it varies across countries and over time.
    Type
    a
  2. Zhu, Y.; Quan, L.; Chen, P.-Y.; Kim, M.C.; Che, C.: Predicting coauthorship using bibliographic network embedding (2023) 0.00
    0.0022374375 = product of:
      0.004474875 = sum of:
        0.004474875 = product of:
          0.00894975 = sum of:
            0.00894975 = weight(_text_:a in 917) [ClassicSimilarity], result of:
              0.00894975 = score(doc=917,freq=14.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.1685276 = fieldWeight in 917, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=917)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Coauthorship prediction applies predictive analytics to bibliographic data to predict authors who are highly likely to be coauthors. In this study, we propose an approach for coauthorship prediction based on bibliographic network embedding through a graph-based bibliographic data model that can be used to model common bibliographic data, including papers, terms, sources, authors, departments, research interests, universities, and countries. A real-world dataset released by AMiner that includes more than 2 million papers, 8 million citations, and 1.7 million authors were integrated into a large bibliographic network using the proposed bibliographic data model. Translation-based methods were applied to the entities and relationships to generate their low-dimensional embeddings while preserving their connectivity information in the original bibliographic network. We applied machine learning algorithms to embeddings that represent the coauthorship relationships of the two authors and achieved high prediction results. The reference model, which is the combination of a network embedding size of 100, the most basic translation-based method, and a gradient boosting method achieved an F1 score of 0.9 and even higher scores are obtainable with different embedding sizes and more advanced embedding methods. Thus, the strengths of the proposed approach lie in its customizable components under a unified framework.
    Type
    a
  3. Wiggers, G.; Verberne, S.; Loon, W. van; Zwenne, G.-J.: Bibliometric-enhanced legal information retrieval : combining usage and citations as flavors of impact relevance (2023) 0.00
    0.0022374375 = product of:
      0.004474875 = sum of:
        0.004474875 = product of:
          0.00894975 = sum of:
            0.00894975 = weight(_text_:a in 1022) [ClassicSimilarity], result of:
              0.00894975 = score(doc=1022,freq=14.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.1685276 = fieldWeight in 1022, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1022)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Bibliometric-enhanced information retrieval uses bibliometrics (e.g., citations) to improve ranking algorithms. Using a data-driven approach, this article describes the development of a bibliometric-enhanced ranking algorithm for legal information retrieval, and the evaluation thereof. We statistically analyze the correlation between usage of documents and citations over time, using data from a commercial legal search engine. We then propose a bibliometric boost function that combines usage of documents with citation counts. The core of this function is an impact variable based on usage and citations that increases in influence as citations and usage counts become more reliable over time. We evaluate our ranking function by comparing search sessions before and after the introduction of the new ranking in the search engine. Using a cost model applied to 129,571 sessions before and 143,864 sessions after the intervention, we show that our bibliometric-enhanced ranking algorithm reduces the time of a search session of legal professionals by 2 to 3% on average for use cases other than known-item retrieval or updating behavior. Given the high hourly tariff of legal professionals and the limited time they can spend on research, this is expected to lead to increased efficiency, especially for users with extremely long search sessions.
    Type
    a
  4. Tian, W.; Cai, R.; Fang, Z.; Geng, Y.; Wang, X.; Hu, Z.: Understanding co-corresponding authorship : a bibliometric analysis and detailed overview (2024) 0.00
    0.0022374375 = product of:
      0.004474875 = sum of:
        0.004474875 = product of:
          0.00894975 = sum of:
            0.00894975 = weight(_text_:a in 1196) [ClassicSimilarity], result of:
              0.00894975 = score(doc=1196,freq=14.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.1685276 = fieldWeight in 1196, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1196)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The phenomenon of co-corresponding authorship is becoming more and more common. To understand the practice of authorship credit sharing among multiple corresponding authors, we comprehensively analyzed the characteristics of the phenomenon of co-corresponding authorships from the perspectives of countries, disciplines, journals, and articles. This researcher was based on a dataset of nearly 8 million articles indexed in the Web of Science, which provides systematic, cross-disciplinary, and large-scale evidence for understanding the phenomenon of co-corresponding authorship for the first time. Our findings reveal that higher proportions of co-corresponding authorship exist in Asian countries, especially in China. From the perspective of disciplines, there is a relatively higher proportion of co-corresponding authorship in the fields of engineering and medicine, while a lower proportion exists in the humanities, social sciences, and computer science fields. From the perspective of journals, high-quality journals usually have higher proportions of co-corresponding authorship. At the level of the article, our findings proved that, compared to articles with a single corresponding author, articles with multiple corresponding authors have a significant citation advantage.
    Type
    a
  5. Costas, R.; Rijcke, S. de; Marres, N.: "Heterogeneous couplings" : operationalizing network perspectives to study science-society interactions through social media metrics (2021) 0.00
    0.0020714647 = product of:
      0.0041429293 = sum of:
        0.0041429293 = product of:
          0.008285859 = sum of:
            0.008285859 = weight(_text_:a in 215) [ClassicSimilarity], result of:
              0.008285859 = score(doc=215,freq=12.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.15602624 = fieldWeight in 215, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=215)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Social media metrics have a genuine networked nature, reflecting the networking characteristics of the social media platform from where they are derived. This networked nature has been relatively less explored in the literature on altmetrics, although new network-level approaches are starting to appear. A general conceptualization of the role of social media networks in science communication, and particularly of social media as a specific type of interface between science and society, is still missing. The aim of this paper is to provide a conceptual framework for appraising interactions between science and society in multiple directions, in what we call heterogeneous couplings. Heterogeneous couplings are conceptualized as the co-occurrence of science and non-science objects, actors, and interactions in online media environments. This conceptualization provides a common framework to study the interactions between science and non-science actors as captured via online and social media platforms. The conceptualization of heterogeneous couplings opens wider opportunities for the development of network applications and analyses of the interactions between societal and scholarly entities in social media environments, paving the way toward more advanced forms of altmetrics, social (media) studies of science, and the conceptualization and operationalization of more advanced science-society studies.
    Type
    a
  6. Zhang, L.; Gou, Z.; Fang, Z.; Sivertsen, G.; Huang, Y.: Who tweets scientific publications? : a large-scale study of tweeting audiences in all areas of research (2023) 0.00
    0.0020714647 = product of:
      0.0041429293 = sum of:
        0.0041429293 = product of:
          0.008285859 = sum of:
            0.008285859 = weight(_text_:a in 1189) [ClassicSimilarity], result of:
              0.008285859 = score(doc=1189,freq=12.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.15602624 = fieldWeight in 1189, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1189)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The purpose of this study is to investigate the validity of tweets about scientific publications as an indicator of societal impact by measuring the degree to which the publications are tweeted beyond academia. We introduce methods that allow for using a much larger and broader data set than in previous validation studies. It covers all areas of research and includes almost 40 million tweets by 2.5 million unique tweeters mentioning almost 4 million scientific publications. We find that, although half of the tweeters are external to academia, most of the tweets are from within academia, and most of the external tweets are responses to original tweets within academia. Only half of the tweeted publications are tweeted outside of academia. We conclude that, in general, the tweeting of scientific publications is not a valid indicator of the societal impact of research. However, publications that continue being tweeted after a few days represent recent scientific achievements that catch attention in society. These publications occur more often in the health sciences and in the social sciences and humanities.
    Content
    Beitrag in: JASIST special issue on 'Who tweets scientific publications? A large-scale study of tweeting audiences in all areas of research'. Vgl.: https://asistdl.onlinelibrary.wiley.com/doi/10.1002/asi.24830.
    Type
    a
  7. Springer, M.: Erlahmt die Forschung? (2023) 0.00
    0.0020296127 = product of:
      0.0040592253 = sum of:
        0.0040592253 = product of:
          0.008118451 = sum of:
            0.008118451 = weight(_text_:a in 985) [ClassicSimilarity], result of:
              0.008118451 = score(doc=985,freq=2.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.15287387 = fieldWeight in 985, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=985)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  8. Liu, X.; Bu, Y.; Li, M.; Li, J.: Monodisciplinary collaboration disrupts science more than multidisciplinary collaboration (2024) 0.00
    0.0020296127 = product of:
      0.0040592253 = sum of:
        0.0040592253 = product of:
          0.008118451 = sum of:
            0.008118451 = weight(_text_:a in 1202) [ClassicSimilarity], result of:
              0.008118451 = score(doc=1202,freq=8.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.15287387 = fieldWeight in 1202, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1202)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Collaboration across disciplines is a critical form of scientific collaboration to solve complex problems and make innovative contributions. This study focuses on the association between multidisciplinary collaboration measured by coauthorship in publications and the disruption of publications measured by the Disruption (D) index. We used authors' affiliations as a proxy of the disciplines to which they belong and categorized an article into multidisciplinary collaboration or monodisciplinary collaboration. The D index quantifies the extent to which a study disrupts its predecessors. We selected 13 journals that publish articles in six disciplines from the Microsoft Academic Graph (MAG) database and then constructed regression models with fixed effects and estimated the relationship between the variables. The findings show that articles with monodisciplinary collaboration are more disruptive than those with multidisciplinary collaboration. Furthermore, we uncovered the mechanism of how monodisciplinary collaboration disrupts science more than multidisciplinary collaboration by exploring the references of the sampled publications.
    Type
    a
  9. Herb, U.; Geith, U.: Kriterien der qualitativen Bewertung wissenschaftlicher Publikationen : Befunde aus dem Projekt visOA (2020) 0.00
    0.001913537 = product of:
      0.003827074 = sum of:
        0.003827074 = product of:
          0.007654148 = sum of:
            0.007654148 = weight(_text_:a in 108) [ClassicSimilarity], result of:
              0.007654148 = score(doc=108,freq=4.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.14413087 = fieldWeight in 108, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=108)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Dieser Beitrag beschreibt a) die Ergebnisse einer Literaturstudie zur qualitativen Wahrnehmung wissenschaftlicher Publikationen, b) die Konstruktion eines daraus abgeleiteten Kriterienkatalogs zur Wahrnehmung der Qualität wissenschaftlicher Publikationen sowie c) der Überprüfung dieses Katalogs in qualitativen Interviews mit Wissenschaflterinnen und Wissenschaftlern aus dem Fachspektrum Chemie, Physik, Biologie, Materialwissenschaft und Engineering. Es zeigte sich, dass die Wahrnehmung von Qualität auf äußerlichen und von außen herangetragenen Faktoren, inhaltlichen / semantischen Faktoren und sprachlichen, syntaktischen sowie strukturellen Faktoren beruht.
    Type
    a
  10. Liu, X.; Chen, X.: Authors' noninstitutional emails and their correlation with retraction (2021) 0.00
    0.001913537 = product of:
      0.003827074 = sum of:
        0.003827074 = product of:
          0.007654148 = sum of:
            0.007654148 = weight(_text_:a in 152) [ClassicSimilarity], result of:
              0.007654148 = score(doc=152,freq=4.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.14413087 = fieldWeight in 152, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=152)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    We collected research articles from Retraction Watch database, Scopus, and a major retraction announcement by Springer, to identify emails used by authors. Authors' emails can be institutional emails and noninstitutional emails. Data suggest that retracted articles are more likely to use noninstitutional emails, but it is difficult to generalize. The study put some focus on authors from China.
    Type
    a
  11. González-Teruel, A.; Pérez-Pulido, M.: ¬The diffusion and influence of theoretical models of information behaviour : the case of Savolainen's ELIS model (2020) 0.00
    0.0018909799 = product of:
      0.0037819599 = sum of:
        0.0037819599 = product of:
          0.0075639198 = sum of:
            0.0075639198 = weight(_text_:a in 5974) [ClassicSimilarity], result of:
              0.0075639198 = score(doc=5974,freq=10.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.14243183 = fieldWeight in 5974, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5974)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    To ascertain the diffusion and influence of Savolainen's ELIS model and its use as a theoretical and/or methodological basis for research. Design/methodology/approach A context citation analysis was made of the work where this researcher published his model. Analysis covered the year of publication, the type of work and the subject matter of the citing documents concerned. In-context citations were analysed for their frequency in each citing text, style, location and content cited. Findings The ELIS model received 18.5 cites/year. 20.2 per cent of them corresponded to papers published in journals in other areas, mainly computer science. The average of cites per paper was 1.8; 64.5 percent of the citing works cited them only once. 60 per cent of the cites were considered essential. Only 13.7 per cent of these cites appear in theory or methods. 37 per cent of the citing documents contained no concept relating to the model. Research limitations/implications The method used focuses on the most direct context of a cite (sentence or paragraph), but isolates it from the general context (full document, other documents by the author or their social capital). It has, however, allowed this research issue to be dealt with under laboratory conditions and revealed nuances hidden by the absolute number of cites. Originality/value It has become evident that the dissemination and influence of the ELIS model are less than what the total number of cites indicates and that it has scarcely been incorporated into research design. Despite its popularity, it is not being validated and/or refuted by way of empirical data.
    Type
    a
  12. Zhang, Y.; Zhang, C.: Enhancing keyphrase extraction from microblogs using human reading time (2021) 0.00
    0.0018909799 = product of:
      0.0037819599 = sum of:
        0.0037819599 = product of:
          0.0075639198 = sum of:
            0.0075639198 = weight(_text_:a in 237) [ClassicSimilarity], result of:
              0.0075639198 = score(doc=237,freq=10.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.14243183 = fieldWeight in 237, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=237)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The premise of manual keyphrase annotation is to read the corresponding content of an annotated object. Intuitively, when we read, more important words will occupy a longer reading time. Hence, by leveraging human reading time, we can find the salient words in the corresponding content. However, previous studies on keyphrase extraction ignore human reading features. In this article, we aim to leverage human reading time to extract keyphrases from microblog posts. There are two main tasks in this study. One is to determine how to measure the time spent by a human on reading a word. We use eye fixation durations (FDs) extracted from an open source eye-tracking corpus. Moreover, we propose strategies to make eye FD more effective on keyphrase extraction. The other task is to determine how to integrate human reading time into keyphrase extraction models. We propose two novel neural network models. The first is a model in which the human reading time is used as the ground truth of the attention mechanism. In the second model, we use human reading time as the external feature. Quantitative and qualitative experiments show that our proposed models yield better performance than the baseline models on two microblog datasets.
    Type
    a
  13. Ikae, C.; Savoy, J.: Gender identification on Twitter (2022) 0.00
    0.0018909799 = product of:
      0.0037819599 = sum of:
        0.0037819599 = product of:
          0.0075639198 = sum of:
            0.0075639198 = weight(_text_:a in 445) [ClassicSimilarity], result of:
              0.0075639198 = score(doc=445,freq=10.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.14243183 = fieldWeight in 445, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=445)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    To determine the author of a text's gender, various feature types have been suggested (e.g., function words, n-gram of letters, etc.) leading to a huge number of stylistic markers. To determine the target category, different machine learning models have been suggested (e.g., logistic regression, decision tree, k nearest-neighbors, support vector machine, naïve Bayes, neural networks, and random forest). In this study, our first objective is to know whether or not the same model always proposes the best effectiveness when considering similar corpora under the same conditions. Thus, based on 7 CLEF-PAN collections, this study analyzes the effectiveness of 10 different classifiers. Our second aim is to propose a 2-stage feature selection to reduce the feature size to a few hundred terms without any significant change in the performance level compared to approaches using all the attributes (increase of around 5% after applying the proposed feature selection). Based on our experiments, neural network or random forest tend, on average, to produce the highest effectiveness. Moreover, empirical evidence indicates that reducing the feature set size to around 300 without penalizing the effectiveness is possible. Finally, based on such reduced feature sizes, an analysis reveals some of the specific terms that clearly discriminate between the 2 genders.
    Type
    a
  14. Liu, M.; Bu, Y.; Chen, C.; Xu, J.; Li, D.; Leng, Y.; Freeman, R.B.; Meyer, E.T.; Yoon, W.; Sung, M.; Jeong, M.; Lee, J.; Kang, J.; Min, C.; Zhai, Y.; Song, M.; Ding, Y.: Pandemics are catalysts of scientific novelty : evidence from COVID-19 (2022) 0.00
    0.0018909799 = product of:
      0.0037819599 = sum of:
        0.0037819599 = product of:
          0.0075639198 = sum of:
            0.0075639198 = weight(_text_:a in 633) [ClassicSimilarity], result of:
              0.0075639198 = score(doc=633,freq=10.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.14243183 = fieldWeight in 633, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=633)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Scientific novelty drives the efforts to invent new vaccines and solutions during the pandemic. First-time collaboration and international collaboration are two pivotal channels to expand teams' search activities for a broader scope of resources required to address the global challenge, which might facilitate the generation of novel ideas. Our analysis of 98,981 coronavirus papers suggests that scientific novelty measured by the BioBERT model that is pretrained on 29 million PubMed articles, and first-time collaboration increased after the outbreak of COVID-19, and international collaboration witnessed a sudden decrease. During COVID-19, papers with more first-time collaboration were found to be more novel and international collaboration did not hamper novelty as it had done in the normal periods. The findings suggest the necessity of reaching out for distant resources and the importance of maintaining a collaborative scientific community beyond nationalism during a pandemic.
    Type
    a
  15. Kulczycki, E.; Huang, Y.; Zuccala, A.A.; Engels, T.C.E.; Ferrara, A.; Guns, R.; Pölönen, J.; Sivertsen, G.; Taskin, Z.; Zhang, L.: Uses of the Journal Impact Factor in national journal rankings in China and Europe (2022) 0.00
    0.0018909799 = product of:
      0.0037819599 = sum of:
        0.0037819599 = product of:
          0.0075639198 = sum of:
            0.0075639198 = weight(_text_:a in 769) [ClassicSimilarity], result of:
              0.0075639198 = score(doc=769,freq=10.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.14243183 = fieldWeight in 769, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=769)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This paper investigates different uses of the Journal Impact Factor (JIF) in national journal rankings and discusses the merits of supplementing metrics with expert assessment. Our focus is national journal rankings used as evidence to support decisions about the distribution of institutional funding or career advancement. The seven countries under comparison are China, Denmark, Finland, Italy, Norway, Poland, and Turkey-and the region of Flanders in Belgium. With the exception of Italy, top-tier journals used in national rankings include those classified at the highest level, or according to tier, or points implemented. A total of 3,565 (75.8%) out of 4,701 unique top-tier journals were identified as having a JIF, with 55.7% belonging to the first Journal Impact Factor quartile. Journal rankings in China, Flanders, Poland, and Turkey classify journals with a JIF as being top-tier, but only when they are in the first quartile of the Average Journal Impact Factor Percentile. Journal rankings that result from expert assessment in Denmark, Finland, and Norway regularly classify journals as top-tier outside the first quartile, particularly in the social sciences and humanities. We conclude that experts, when tasked with metric-informed journal rankings, take into account quality dimensions that are not covered by JIFs.
    Type
    a
  16. Cui, Y.; Wang, Y.; Liu, X.; Wang, X.; Zhang, X.: Multidimensional scholarly citations : characterizing and understanding scholars' citation behaviors (2023) 0.00
    0.0018909799 = product of:
      0.0037819599 = sum of:
        0.0037819599 = product of:
          0.0075639198 = sum of:
            0.0075639198 = weight(_text_:a in 847) [ClassicSimilarity], result of:
              0.0075639198 = score(doc=847,freq=10.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.14243183 = fieldWeight in 847, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=847)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This study investigates scholars' citation behaviors from a fine-grained perspective. Specifically, each scholarly citation is considered multidimensional rather than logically unidimensional (i.e., present or absent). Thirty million articles from PubMed were accessed for use in empirical research, in which a total of 15 interpretable features of scholarly citations were constructed and grouped into three main categories. Each category corresponds to one aspect of the reasons and motivations behind scholars' citation decision-making during academic writing. Using about 500,000 pairs of actual and randomly generated scholarly citations, a series of Random Forest-based classification experiments were conducted to quantitatively evaluate the correlation between each constructed citation feature and citation decisions made by scholars. Our experimental results indicate that citation proximity is the category most relevant to scholars' citation decision-making, followed by citation authority and citation inertia. However, big-name scholars whose h-indexes rank among the top 1% exhibit a unique pattern of citation behaviors-their citation decision-making correlates most closely with citation inertia, with the correlation nearly three times as strong as that of their ordinary counterparts. Hopefully, the empirical findings presented in this paper can bring us closer to characterizing and understanding the complex process of generating scholarly citations in academia.
    Type
    a
  17. Zhao, D.; Strotmann, A.: Intellectual structure of information science 2011-2020 : an author co-citation analysis (2022) 0.00
    0.0017899501 = product of:
      0.0035799001 = sum of:
        0.0035799001 = product of:
          0.0071598003 = sum of:
            0.0071598003 = weight(_text_:a in 610) [ClassicSimilarity], result of:
              0.0071598003 = score(doc=610,freq=14.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.13482209 = fieldWeight in 610, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.03125 = fieldNorm(doc=610)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Purpose This study continues a long history of author co-citation analysis of the intellectual structure of information science into the time period of 2011-2020. It also examines changes in this structure from 2006-2010 through 2011-2015 to 2016-2020. Results will contribute to a better understanding of the information science research field. Design/methodology/approach The well-established procedures and techniques for author co-citation analysis were followed. Full records of research articles in core information science journals published during 2011-2020 were retrieved and downloaded from the Web of Science database. About 150 most highly cited authors in each of the two five-year time periods were selected from this dataset to represent this field, and their co-citation counts were calculated. Each co-citation matrix was input into SPSS for factor analysis, and results were visualized in Pajek. Factors were interpreted as specialties and labeled upon an examination of articles written by authors who load primarily on each factor. Findings The two-camp structure of information science continued to be present clearly. Bibliometric indicators for research evaluation dominated the Knowledge Domain Analysis camp during both fivr-year time periods, whereas interactive information retrieval (IR) dominated the IR camp during 2011-2015 but shared dominance with information behavior during 2016-2020. Bridging between the two camps became increasingly weaker and was only provided by the scholarly communication specialty during 2016-2020. The IR systems specialty drifted further away from the IR camp. The information behavior specialty experienced a deep slump during 2011-2020 in its evolution process. Altmetrics grew to dominate the Webometrics specialty and brought it to a sharp increase during 2016-2020. Originality/value Author co-citation analysis (ACA) is effective in revealing intellectual structures of research fields. Most related studies used term-based methods to identify individual research topics but did not examine the interrelationships between these topics or the overall structure of the field. The few studies that did discuss the overall structure paid little attention to the effect of changes to the source journals on the results. The present study does not have these problems and continues the long history of benchmark contributions to a better understanding of the information science field using ACA.
    Type
    a
  18. Tay, W.; Zhang, X.; Karimi , S.: Beyond mean rating : probabilistic aggregation of star ratings based on helpfulness (2020) 0.00
    0.001757696 = product of:
      0.003515392 = sum of:
        0.003515392 = product of:
          0.007030784 = sum of:
            0.007030784 = weight(_text_:a in 5917) [ClassicSimilarity], result of:
              0.007030784 = score(doc=5917,freq=6.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.13239266 = fieldWeight in 5917, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5917)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The star-rating mechanism of customer reviews is used universally by the online population to compare and select merchants, movies, products, and services. The consensus opinion from aggregation of star ratings is used as a proxy for item quality. Online reviews are noisy and effective aggregation of star ratings to accurately reflect the "true quality" of products and services is challenging. The mean-rating aggregation model is widely used and other aggregation models are also proposed. These existing aggregation models rely on a large number of reviews to tolerate noise. However, many products rarely have reviews. We propose probabilistic aggregation models for review ratings based on the Dirichlet distribution to combat data sparsity in reviews. We further propose to exploit the "helpfulness" social information and time to filter noisy reviews and effectively aggregate ratings to compute the consensus opinion. Our experiments on an Amazon data set show that our probabilistic aggregation models based on "helpfulness" achieve better performance than the statistical and heuristic baseline approaches.
    Type
    a
  19. Wu, C.; Yan, E.; Zhu, Y.; Li, K.: Gender imbalance in the productivity of funded projects : a study of the outputs of National Institutes of Health R01 grants (2021) 0.00
    0.001757696 = product of:
      0.003515392 = sum of:
        0.003515392 = product of:
          0.007030784 = sum of:
            0.007030784 = weight(_text_:a in 391) [ClassicSimilarity], result of:
              0.007030784 = score(doc=391,freq=6.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.13239266 = fieldWeight in 391, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=391)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This study examines the relationship between team's gender composition and outputs of funded projects using a large data set of National Institutes of Health (NIH) R01 grants and their associated publications between 1990 and 2017. This study finds that while the women investigators' presence in NIH grants is generally low, higher women investigator presence is on average related to slightly lower number of publications. This study finds empirically that women investigators elect to work in fields in which fewer publications per million-dollar funding is the norm. For fields where women investigators are relatively well represented, they are as productive as men. The overall lower productivity of women investigators may be attributed to the low representation of women in high productivity fields dominated by men investigators. The findings shed light on possible reasons for gender disparity in grant productivity.
    Type
    a
  20. Fang, Z.; Dudek, J.; Costas, R.: Facing the volatility of tweets in altmetric research (2022) 0.00
    0.001757696 = product of:
      0.003515392 = sum of:
        0.003515392 = product of:
          0.007030784 = sum of:
            0.007030784 = weight(_text_:a in 605) [ClassicSimilarity], result of:
              0.007030784 = score(doc=605,freq=6.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.13239266 = fieldWeight in 605, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=605)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The data re-collection for tweets from data snapshots is a common methodological step in Twitter-based research. Understanding better the volatility of tweets over time is important for validating the reliability of metrics based on Twitter data. We tracked a set of 37,918 original scholarly tweets mentioning COVID-19-related research daily for 56 days and captured the reasons for the changes in their availability over time. Results show that the proportion of unavailable tweets increased from 1.6 to 2.6% in the time window observed. Of the 1,323 tweets that became unavailable at some point in the period observed, 30.5% became available again afterwards. "Revived" tweets resulted mainly from the unprotecting, reactivating, or unsuspending of users' accounts. Our findings highlight the importance of noting this dynamic nature of Twitter data in altmetric research and testify to the challenges that this poses for the retrieval, processing, and interpretation of Twitter data about scientific papers.
    Type
    a