Search (44 results, page 1 of 3)

  • × author_ss:"Bornmann, L."
  • × theme_ss:"Informetrie"
  1. Bornmann, L.; Haunschild, R.: Relative Citation Ratio (RCR) : an empirical attempt to study a new field-normalized bibliometric indicator (2017) 0.04
    0.040197317 = product of:
      0.10049329 = sum of:
        0.007388207 = weight(_text_:a in 3541) [ClassicSimilarity], result of:
          0.007388207 = score(doc=3541,freq=6.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.1544581 = fieldWeight in 3541, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3541)
        0.093105085 = weight(_text_:68 in 3541) [ClassicSimilarity], result of:
          0.093105085 = score(doc=3541,freq=2.0), product of:
            0.2234734 = queryWeight, product of:
              5.386969 = idf(docFreq=549, maxDocs=44218)
              0.04148407 = queryNorm
            0.41662714 = fieldWeight in 3541, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.386969 = idf(docFreq=549, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3541)
      0.4 = coord(2/5)
    
    Abstract
    Hutchins, Yuan, Anderson, and Santangelo (2015) proposed the Relative Citation Ratio (RCR) as a new field-normalized impact indicator. This study investigates the RCR by correlating it on the level of single publications with established field-normalized indicators and assessments of the publications by peers. We find that the RCR correlates highly with established field-normalized indicators, but the correlation between RCR and peer assessments is only low to medium.
    Source
    Journal of the Association for Information Science and Technology. 68(2017) no.4, S.1064-1067
    Type
    a
  2. Bornmann, L.; Haunschild, R.: ¬An empirical look at the nature index (2017) 0.03
    0.028325006 = product of:
      0.070812516 = sum of:
        0.00430889 = weight(_text_:a in 3432) [ClassicSimilarity], result of:
          0.00430889 = score(doc=3432,freq=4.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.090081796 = fieldWeight in 3432, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3432)
        0.06650363 = weight(_text_:68 in 3432) [ClassicSimilarity], result of:
          0.06650363 = score(doc=3432,freq=2.0), product of:
            0.2234734 = queryWeight, product of:
              5.386969 = idf(docFreq=549, maxDocs=44218)
              0.04148407 = queryNorm
            0.29759082 = fieldWeight in 3432, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.386969 = idf(docFreq=549, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3432)
      0.4 = coord(2/5)
    
    Abstract
    In November 2014, the Nature Index (NI) was introduced (see http://www.natureindex.com) by the Nature Publishing Group (NPG). The NI comprises the primary research articles published in the past 12 months in a selection of reputable journals. Starting from two short comments on the NI (Haunschild & Bornmann, 2015a, 2015b), we undertake an empirical analysis of the NI using comprehensive country data. We investigate whether the huge efforts of computing the NI are justified and whether the size-dependent NI indicators should be complemented by size-independent variants. The analysis uses data from the Max Planck Digital Library in-house database (which is based on Web of Science data) and from the NPG. In the first step of the analysis, we correlate the NI with other metrics that are simpler to generate than the NI. The resulting large correlation coefficients point out that the NI produces similar results as simpler solutions. In the second step of the analysis, relative and size-independent variants of the NI are generated that should be additionally presented by the NPG. The size-dependent NI indicators favor large countries (or institutions) and the top-performing small countries (or institutions) do not come into the picture.
    Source
    Journal of the Association for Information Science and Technology. 68(2017) no.3, S.653-659
    Type
    a
  3. Bornmann, L.: Is collaboration among scientists related to the citation impact of papers because their quality increases with collaboration? : an analysis based on data from F1000Prime and normalized citation scores (2017) 0.03
    0.02782019 = product of:
      0.06955048 = sum of:
        0.0030468449 = weight(_text_:a in 3539) [ClassicSimilarity], result of:
          0.0030468449 = score(doc=3539,freq=2.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.06369744 = fieldWeight in 3539, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3539)
        0.06650363 = weight(_text_:68 in 3539) [ClassicSimilarity], result of:
          0.06650363 = score(doc=3539,freq=2.0), product of:
            0.2234734 = queryWeight, product of:
              5.386969 = idf(docFreq=549, maxDocs=44218)
              0.04148407 = queryNorm
            0.29759082 = fieldWeight in 3539, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.386969 = idf(docFreq=549, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3539)
      0.4 = coord(2/5)
    
    Source
    Journal of the Association for Information Science and Technology. 68(2017) no.4, S.1036-1047
    Type
    a
  4. Marx, W.; Bornmann, L.: On the problems of dealing with bibliometric data (2014) 0.02
    0.016414216 = product of:
      0.041035537 = sum of:
        0.007312428 = weight(_text_:a in 1239) [ClassicSimilarity], result of:
          0.007312428 = score(doc=1239,freq=2.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.15287387 = fieldWeight in 1239, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.09375 = fieldNorm(doc=1239)
        0.03372311 = product of:
          0.06744622 = sum of:
            0.06744622 = weight(_text_:22 in 1239) [ClassicSimilarity], result of:
              0.06744622 = score(doc=1239,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.46428138 = fieldWeight in 1239, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=1239)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Date
    18. 3.2014 19:13:22
    Type
    a
  5. Bornmann, L.; Mutz, R.: From P100 to P100' : a new citation-rank approach (2014) 0.01
    0.012892792 = product of:
      0.03223198 = sum of:
        0.009749904 = weight(_text_:a in 1431) [ClassicSimilarity], result of:
          0.009749904 = score(doc=1431,freq=8.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.20383182 = fieldWeight in 1431, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=1431)
        0.022482075 = product of:
          0.04496415 = sum of:
            0.04496415 = weight(_text_:22 in 1431) [ClassicSimilarity], result of:
              0.04496415 = score(doc=1431,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.30952093 = fieldWeight in 1431, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1431)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Properties of a percentile-based rating scale needed in bibliometrics are formulated. Based on these properties, P100 was recently introduced as a new citation-rank approach (Bornmann, Leydesdorff, & Wang, 2013). In this paper, we conceptualize P100 and propose an improvement which we call P100'. Advantages and disadvantages of citation-rank indicators are noted.
    Date
    22. 8.2014 17:05:18
    Type
    a
  6. Leydesdorff, L.; Bornmann, L.; Wagner, C.S.: ¬The relative influences of government funding and international collaboration on citation impact (2019) 0.01
    0.0100148395 = product of:
      0.025037099 = sum of:
        0.008175544 = weight(_text_:a in 4681) [ClassicSimilarity], result of:
          0.008175544 = score(doc=4681,freq=10.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.1709182 = fieldWeight in 4681, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=4681)
        0.016861554 = product of:
          0.03372311 = sum of:
            0.03372311 = weight(_text_:22 in 4681) [ClassicSimilarity], result of:
              0.03372311 = score(doc=4681,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.23214069 = fieldWeight in 4681, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4681)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    A recent publication in Nature reports that public R&D funding is only weakly correlated with the citation impact of a nation's articles as measured by the field-weighted citation index (FWCI; defined by Scopus). On the basis of the supplementary data, we up-scaled the design using Web of Science data for the decade 2003-2013 and OECD funding data for the corresponding decade assuming a 2-year delay (2001-2011). Using negative binomial regression analysis, we found very small coefficients, but the effects of international collaboration are positive and statistically significant, whereas the effects of government funding are negative, an order of magnitude smaller, and statistically nonsignificant (in two of three analyses). In other words, international collaboration improves the impact of research articles, whereas more government funding tends to have a small adverse effect when comparing OECD countries.
    Date
    8. 1.2019 18:22:45
    Type
    a
  7. Leydesdorff, L.; Bornmann, L.: How fractional counting of citations affects the impact factor : normalization in terms of differences in citation potentials among fields of science (2011) 0.01
    0.0083457 = product of:
      0.02086425 = sum of:
        0.006812953 = weight(_text_:a in 4186) [ClassicSimilarity], result of:
          0.006812953 = score(doc=4186,freq=10.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.14243183 = fieldWeight in 4186, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4186)
        0.014051297 = product of:
          0.028102593 = sum of:
            0.028102593 = weight(_text_:22 in 4186) [ClassicSimilarity], result of:
              0.028102593 = score(doc=4186,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.19345059 = fieldWeight in 4186, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4186)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    The Impact Factors (IFs) of the Institute for Scientific Information suffer from a number of drawbacks, among them the statistics-Why should one use the mean and not the median?-and the incomparability among fields of science because of systematic differences in citation behavior among fields. Can these drawbacks be counteracted by fractionally counting citation weights instead of using whole numbers in the numerators? (a) Fractional citation counts are normalized in terms of the citing sources and thus would take into account differences in citation behavior among fields of science. (b) Differences in the resulting distributions can be tested statistically for their significance at different levels of aggregation. (c) Fractional counting can be generalized to any document set including journals or groups of journals, and thus the significance of differences among both small and large sets can be tested. A list of fractionally counted IFs for 2008 is available online at http:www.leydesdorff.net/weighted_if/weighted_if.xls The between-group variance among the 13 fields of science identified in the U.S. Science and Engineering Indicators is no longer statistically significant after this normalization. Although citation behavior differs largely between disciplines, the reflection of these differences in fractionally counted citation distributions can not be used as a reliable instrument for the classification.
    Date
    22. 1.2011 12:51:07
    Type
    a
  8. Bornmann, L.: How to analyze percentile citation impact data meaningfully in bibliometrics : the statistical analysis of distributions, percentile rank classes, and top-cited papers (2013) 0.01
    0.008207108 = product of:
      0.020517768 = sum of:
        0.003656214 = weight(_text_:a in 656) [ClassicSimilarity], result of:
          0.003656214 = score(doc=656,freq=2.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.07643694 = fieldWeight in 656, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=656)
        0.016861554 = product of:
          0.03372311 = sum of:
            0.03372311 = weight(_text_:22 in 656) [ClassicSimilarity], result of:
              0.03372311 = score(doc=656,freq=2.0), product of:
                0.14527014 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04148407 = queryNorm
                0.23214069 = fieldWeight in 656, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=656)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Date
    22. 3.2013 19:44:17
    Type
    a
  9. Bornmann, L.; Marx, W.: ¬The wisdom of citing scientists (2014) 0.00
    0.0020897004 = product of:
      0.010448502 = sum of:
        0.010448502 = weight(_text_:a in 1293) [ClassicSimilarity], result of:
          0.010448502 = score(doc=1293,freq=12.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.21843673 = fieldWeight in 1293, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1293)
      0.2 = coord(1/5)
    
    Abstract
    This Brief Communication discusses the benefits of citation analysis in research evaluation based on Galton's "Wisdom of Crowds" (1907). Citations are based on the assessment of many which is why they can be considered to have some credibility. However, we show that citations are incomplete assessments and that one cannot assume that a high number of citations correlates with a high level of usefulness. Only when one knows that a rarely cited paper has been widely read is it possible to say-strictly speaking-that it was obviously of little use for further research. Using a comparison with "like" data, we try to determine that cited reference analysis allows for a more meaningful analysis of bibliometric data than times-cited analysis.
    Type
    a
  10. Bornmann, L.: Is there currently a scientific revolution in Scientometrics? (2014) 0.00
    0.002068267 = product of:
      0.010341335 = sum of:
        0.010341335 = weight(_text_:a in 1206) [ClassicSimilarity], result of:
          0.010341335 = score(doc=1206,freq=4.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.2161963 = fieldWeight in 1206, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.09375 = fieldNorm(doc=1206)
      0.2 = coord(1/5)
    
    Type
    a
  11. Bornmann, L.: What do altmetrics counts mean? : a plea for content analyses (2016) 0.00
    0.002068267 = product of:
      0.010341335 = sum of:
        0.010341335 = weight(_text_:a in 2858) [ClassicSimilarity], result of:
          0.010341335 = score(doc=2858,freq=4.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.2161963 = fieldWeight in 2858, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.09375 = fieldNorm(doc=2858)
      0.2 = coord(1/5)
    
    Type
    a
  12. Bornmann, L.; Bauer, J.: Which of the world's institutions employ the most highly cited researchers : an analysis of the data from highlycited.com (2015) 0.00
    0.0019499809 = product of:
      0.009749904 = sum of:
        0.009749904 = weight(_text_:a in 1556) [ClassicSimilarity], result of:
          0.009749904 = score(doc=1556,freq=8.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.20383182 = fieldWeight in 1556, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=1556)
      0.2 = coord(1/5)
    
    Abstract
    In 2014, Thomson Reuters published a list of the most highly cited researchers worldwide (highlycited.com). Because the data are freely available for downloading and include the names of the researchers' institutions, we produced a ranking of the institutions on the basis of the number of highly cited researchers per institution. This ranking is intended to be a helpful amendment of other available institutional rankings.
    Type
    a
  13. Bornmann, L.; Bauer, J.: Which of the world's institutions employ the most highly cited researchers : an analysis of the data from highlycited.com (2015) 0.00
    0.0019499809 = product of:
      0.009749904 = sum of:
        0.009749904 = weight(_text_:a in 2223) [ClassicSimilarity], result of:
          0.009749904 = score(doc=2223,freq=8.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.20383182 = fieldWeight in 2223, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=2223)
      0.2 = coord(1/5)
    
    Abstract
    In 2014, Thomson Reuters published a list of the most highly cited researchers worldwide (highlycited.com). Because the data are freely available for downloading and include the names of the researchers' institutions, we produced a ranking of the institutions on the basis of the number of highly cited researchers per institution. This ranking is intended to be a helpful amendment of other available institutional rankings.
    Type
    a
  14. Bornmann, L.; Leydesdorff, L.: Which cities produce more excellent papers than can be expected? : a new mapping approach, using Google Maps, based on statistical significance testing (2011) 0.00
    0.0019346869 = product of:
      0.009673434 = sum of:
        0.009673434 = weight(_text_:a in 4767) [ClassicSimilarity], result of:
          0.009673434 = score(doc=4767,freq=14.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.20223314 = fieldWeight in 4767, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=4767)
      0.2 = coord(1/5)
    
    Abstract
    The methods presented in this paper allow for a statistical analysis revealing centers of excellence around the world using programs that are freely available. Based on Web of Science data (a fee-based database), field-specific excellence can be identified in cities where highly cited papers were published more frequently than can be expected. Compared to the mapping approaches published hitherto, our approach is more analytically oriented by allowing the assessment of an observed number of excellent papers for a city against the expected number. Top performers in output are cities in which authors are located who publish a statistically significant higher number of highly cited papers than can be expected for these cities. As sample data for physics, chemistry, and psychology show, these cities do not necessarily have a high output of highly cited papers.
    Type
    a
  15. Bornmann, L.; Daniel, H.-D.: What do we know about the h index? (2007) 0.00
    0.001907627 = product of:
      0.009538135 = sum of:
        0.009538135 = weight(_text_:a in 477) [ClassicSimilarity], result of:
          0.009538135 = score(doc=477,freq=10.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.19940455 = fieldWeight in 477, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=477)
      0.2 = coord(1/5)
    
    Abstract
    Jorge Hirsch recently proposed the h index to quantify the research output of individual scientists. The new index has attracted a lot of attention in the scientific community. The claim that the h index in a single number provides a good representation of the scientific lifetime achievement of a scientist as well as the (supposed) simple calculation of the h index using common literature databases lead to the danger of improper use of the index. We describe the advantages and disadvantages of the h index and summarize the studies on the convergent validity of this index. We also introduce corrections and complements as well as single-number alternatives to the h index.
    Type
    a
  16. Bornmann, L.; Daniel, H.-D.: Universality of citation distributions : a validation of Radicchi et al.'s relative indicator cf = c/c0 at the micro level using data from chemistry (2009) 0.00
    0.001828107 = product of:
      0.009140535 = sum of:
        0.009140535 = weight(_text_:a in 2954) [ClassicSimilarity], result of:
          0.009140535 = score(doc=2954,freq=18.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.19109234 = fieldWeight in 2954, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2954)
      0.2 = coord(1/5)
    
    Abstract
    In a recently published PNAS paper, Radicchi, Fortunato, and Castellano (2008) propose the relative indicator cf as an unbiased indicator for citation performance across disciplines (fields, subject areas). To calculate cf, the citation rate for a single paper is divided by the average number of citations for all papers in the discipline in which the single paper has been categorized. cf values are said to lead to a universality of discipline-specific citation distributions. Using a comprehensive dataset of an evaluation study on Angewandte Chemie International Edition (AC-IE), we tested the advantage of using this indicator in practical application at the micro level, as compared with (1) simple citation rates, and (2) z-scores, which have been used in psychological testing for many years for normalization of test scores. To calculate z-scores, the mean number of citations of the papers within a discipline is subtracted from the citation rate of a single paper, and the difference is then divided by the citations' standard deviation for a discipline. Our results indicate that z-scores are better suited than cf values to produce universality of discipline-specific citation distributions.
    Type
    a
  17. Bornmann, L.; Mutz, R.; Daniel, H.D.: Do we need the h index and its variants in addition to standard bibliometric measures? (2009) 0.00
    0.0017235561 = product of:
      0.00861778 = sum of:
        0.00861778 = weight(_text_:a in 2861) [ClassicSimilarity], result of:
          0.00861778 = score(doc=2861,freq=16.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.18016359 = fieldWeight in 2861, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2861)
      0.2 = coord(1/5)
    
    Abstract
    In this study, we investigate whether there is a need for the h index and its variants in addition to standard bibliometric measures (SBMs). Results from our recent study (L. Bornmann, R. Mutz, & H.-D. Daniel, 2008) have indicated that there are two types of indices: One type of indices (e.g., h index) describes the most productive core of a scientist's output and informs about the number of papers in the core. The other type of indices (e.g., a index) depicts the impact of the papers in the core. In evaluative bibliometric studies, the two dimensions quantity and quality of output are usually assessed using the SBMs number of publications (for the quantity dimension) and total citation counts (for the impact dimension). We additionally included the SBMs into the factor analysis. The results of the newly calculated analysis indicate that there is a high intercorrelation between number of publications and the indices that load substantially on the factor Quantity of the Productive Core as well as between total citation counts and the indices that load substantially on the factor Impact of the Productive Core. The high-loading indices and SBMs within one performance dimension could be called redundant in empirical application, as high intercorrelations between different indicators are a sign for measuring something similar (or the same). Based on our findings, we propose the use of any pair of indicators (one relating to the number of papers in a researcher's productive core and one relating to the impact of these core papers) as a meaningful approach for comparing scientists.
    Type
    a
  18. Bornmann, L.: How much does the expected number of citations for a publication change if it contains the address of a specific scientific institute? : a new approach for the analysis of citation data on the institutional level based on regression models (2016) 0.00
    0.0017235561 = product of:
      0.00861778 = sum of:
        0.00861778 = weight(_text_:a in 3095) [ClassicSimilarity], result of:
          0.00861778 = score(doc=3095,freq=16.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.18016359 = fieldWeight in 3095, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3095)
      0.2 = coord(1/5)
    
    Abstract
    Citation data for institutes are generally provided as numbers of citations or as relative citation rates (as, for example, in the Leiden Ranking). These numbers can then be compared between the institutes. This study aims to present a new approach for the evaluation of citation data at the institutional level, based on regression models. As example data, the study includes all articles and reviews from the Web of Science for the publication year 2003 (n?=?886,416 papers). The study is based on an in-house database of the Max Planck Society. The study investigates how much the expected number of citations for a publication changes if it contains the address of an institute. The calculation of the expected values allows, on the one hand, investigating how the citation impact of the papers of an institute appears in comparison with the total of all papers. On the other hand, the expected values for several institutes can be compared with one another or with a set of randomly selected publications. Besides the institutes, the regression models include factors which can be assumed to have a general influence on citation counts (e.g., the number of authors).
    Type
    a
  19. Bornmann, L.: How well does a university perform in comparison with its peers? : The use of odds, and odds ratios, for the comparison of institutional citation impact using the Leiden Rankings (2015) 0.00
    0.0017062332 = product of:
      0.008531166 = sum of:
        0.008531166 = weight(_text_:a in 2340) [ClassicSimilarity], result of:
          0.008531166 = score(doc=2340,freq=8.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.17835285 = fieldWeight in 2340, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2340)
      0.2 = coord(1/5)
    
    Abstract
    This study presents the calculation of odds, and odds ratios, for the comparison of the citation impact of universities in the Leiden Ranking. Odds and odds ratios can be used to measure the performance difference between a selected university and competing institutions, or the average of selected competitors, in a relatively simple but clear way.
    Type
    a
  20. Bornmann, L.; Daniel, H.-D.: Multiple publication on a single research study: does it pay? : The influence of number of research articles on total citation counts in biomedicine (2007) 0.00
    0.0016122389 = product of:
      0.008061195 = sum of:
        0.008061195 = weight(_text_:a in 444) [ClassicSimilarity], result of:
          0.008061195 = score(doc=444,freq=14.0), product of:
            0.04783308 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.04148407 = queryNorm
            0.1685276 = fieldWeight in 444, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=444)
      0.2 = coord(1/5)
    
    Abstract
    Scientists may seek to report a single definable body of research in more than one publication, that is, in repeated reports of the same work or in fractional reports, in order to disseminate their research as widely as possible in the scientific community. Up to now, however, it has not been examined whether this strategy of "multiple publication" in fact leads to greater reception of the research. In the present study, we investigate the influence of number of articles reporting the results of a single study on reception in the scientific community (total citation counts of an article on a single study). Our data set consists of 96 applicants for a research fellowship from the Boehringer Ingelheim Fonds (BIF), an international foundation for the promotion of basic research in biomedicine. The applicants reported to us all articles that they had published within the framework of their doctoral research projects. On this single project, the applicants had published from 1 to 16 articles (M = 4; Mdn = 3). The results of a regression model with an interaction term show that the practice of multiple publication of research study results does in fact lead to greater reception of the research (higher total citation counts) in the scientific community. However, reception is dependent upon length of article: the longer the article, the more total citation counts increase with the number of articles. Thus, it pays for scientists to practice multiple publication of study results in the form of sizable reports.
    Type
    a