Search (6 results, page 1 of 1)

  • × author_ss:"D'Angelo, C.A."
  • × theme_ss:"Informetrie"
  1. Abramo, G.; D'Angelo, C.A.; Viel, F.: Assessing the accuracy of the h- and g-indexes for measuring researchers' productivity (2013) 0.02
    0.024069263 = product of:
      0.12034631 = sum of:
        0.12034631 = weight(_text_:index in 957) [ClassicSimilarity], result of:
          0.12034631 = score(doc=957,freq=10.0), product of:
            0.18579477 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.04251826 = queryNorm
            0.64773786 = fieldWeight in 957, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.046875 = fieldNorm(doc=957)
      0.2 = coord(1/5)
    
    Abstract
    Bibliometric indicators are increasingly used in support of decisions about recruitment, career advancement, rewards, and selective funding for scientists. Given the importance of the applications, bibliometricians are obligated to carry out empirical testing of the robustness of the indicators, in simulations of real contexts. In this work, we compare the results of national-scale research assessments at the individual level, based on the following three different indexes: the h-index, the g-index, and "fractional scientific strength" (FSS), an indicator previously proposed by the authors. For each index, we construct and compare rankings lists of all Italian academic researchers working in the hard sciences during the period 2001-2005. The analysis quantifies the shifts in ranks that occur when researchers' productivity rankings by simple indicators such as the h- or g-indexes are compared with those by more accurate FSS.
    Object
    g-index
    h-index
  2. Abramo, G.; D'Angelo, C.A.; Viel, F.: ¬A robust benchmark for the h- and g-indexes (2010) 0.02
    0.01864397 = product of:
      0.09321985 = sum of:
        0.09321985 = weight(_text_:index in 3470) [ClassicSimilarity], result of:
          0.09321985 = score(doc=3470,freq=6.0), product of:
            0.18579477 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.04251826 = queryNorm
            0.50173557 = fieldWeight in 3470, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.046875 = fieldNorm(doc=3470)
      0.2 = coord(1/5)
    
    Abstract
    The use of Hirsch's h-index as a joint proxy of the impact and productivity of a scientist's research work continues to gain ground, accompanied by the efforts of bibliometrists to resolve some of its critical issues through the application of a number of more or less sophisticated variants. However, the literature does not reveal any appreciable attempt to overcome the objective problems of measuring h-indexes on a large scale for purposes of comparative evaluation. Scientists may succeed in calculating their own h-indexes but, being unable to compare them to those of their peers, they are unable to obtain truly useful indications of their individual research performance. This study proposes to overcome this gap, measuring the h- and Egghe's g-indexes of all Italian university researchers in the hard sciences over a 5-year window. Descriptive statistics are provided concerning all of the 165 subject fields examined, offering robust benchmarks for those who wish to compare their individual performance to those of their colleagues in the same subject field.
    Object
    h-index
    g-index
  3. D'Angelo, C.A.; Giuffrida, C.; Abramo, G.: ¬A heuristic approach to author name disambiguation in bibliometrics databases for large-scale research assessments (2011) 0.02
    0.015792316 = product of:
      0.039480787 = sum of:
        0.027959513 = weight(_text_:system in 4190) [ClassicSimilarity], result of:
          0.027959513 = score(doc=4190,freq=2.0), product of:
            0.13391352 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.04251826 = queryNorm
            0.20878783 = fieldWeight in 4190, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.046875 = fieldNorm(doc=4190)
        0.011521274 = product of:
          0.03456382 = sum of:
            0.03456382 = weight(_text_:22 in 4190) [ClassicSimilarity], result of:
              0.03456382 = score(doc=4190,freq=2.0), product of:
                0.1488917 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04251826 = queryNorm
                0.23214069 = fieldWeight in 4190, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4190)
          0.33333334 = coord(1/3)
      0.4 = coord(2/5)
    
    Abstract
    National exercises for the evaluation of research activity by universities are becoming regular practice in ever more countries. These exercises have mainly been conducted through the application of peer-review methods. Bibliometrics has not been able to offer a valid large-scale alternative because of almost overwhelming difficulties in identifying the true author of each publication. We will address this problem by presenting a heuristic approach to author name disambiguation in bibliometric datasets for large-scale research assessments. The application proposed concerns the Italian university system, comprising 80 universities and a research staff of over 60,000 scientists. The key advantage of the proposed approach is the ease of implementation. The algorithms are of practical application and have considerably better scalability and expandability properties than state-of-the-art unsupervised approaches. Moreover, the performance in terms of precision and recall, which can be further improved, seems thoroughly adequate for the typical needs of large-scale bibliometric research assessments.
    Date
    22. 1.2011 13:06:52
  4. Abramo, G.; D'Angelo, C.A.; Costa, F. Di: Identifying interdisciplinarity through the disciplinary classification of coauthors of scientific publications (2012) 0.01
    0.0107641 = product of:
      0.0538205 = sum of:
        0.0538205 = weight(_text_:index in 491) [ClassicSimilarity], result of:
          0.0538205 = score(doc=491,freq=2.0), product of:
            0.18579477 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.04251826 = queryNorm
            0.28967714 = fieldWeight in 491, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.046875 = fieldNorm(doc=491)
      0.2 = coord(1/5)
    
    Abstract
    The growing complexity of challenges involved in scientific progress demands ever more frequent application of competencies and knowledge from different scientific fields. The present work analyzes the degree of collaboration among scientists from different disciplines to identify the most frequent "combinations of knowledge" in research activity. The methodology adopts an innovative bibliometric approach based on the disciplinary affiliation of publication coauthors. The field of observation includes all publications (167,179) indexed in the Science Citation Index Expanded for the years 2004-2008, authored by all scientists in the hard sciences (43,223) at Italian universities (68). The analysis examines 205 research fields grouped in 9 disciplines. Identifying the fields with the highest potential of interdisciplinary collaboration is useful to inform research polices at the national and regional levels, as well as management strategies at the institutional level.
  5. Abramo, G.; D'Angelo, C.A.: ¬A decision support system for public research organizations participating in national research assessment exercises (2009) 0.01
    0.0065901205 = product of:
      0.032950602 = sum of:
        0.032950602 = weight(_text_:system in 3123) [ClassicSimilarity], result of:
          0.032950602 = score(doc=3123,freq=4.0), product of:
            0.13391352 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.04251826 = queryNorm
            0.24605882 = fieldWeight in 3123, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3123)
      0.2 = coord(1/5)
    
    Abstract
    We are witnessing a rapid trend toward the adoption of exercises for evaluation of national research systems, generally based on peer review. They respond to two main needs: stimulating higher efficiency in research activities by public laboratories, and realizing better allocative efficiency in government funding of such institutions. However, the peer review approach is typified by several limitations that raise doubts for the achievement of the ultimate objectives. In particular, subjectivity of judgment, which occurs during the step of selecting research outputs to be submitted for the evaluations, risks heavily distorting both the final ratings of the organizations evaluated and the ultimate funding they receive. These distortions become ever more relevant if the evaluation is limited to small samples of the scientific production of the research institutions. The objective of the current study is to propose a quantitative methodology based on bibliometric data that would provide a reliable support for the process of selecting the best products of a laboratory, and thus limit distortions. Benefits are twofold: single research institutions can maximize the probability of receiving a fair evaluation coherent with the real quality of their research. At the same time, broader adoptions of this approach could also provide strong advantages at the macroeconomic level, since it guarantees financial allocations based on the real value of the institutions under evaluation. In this study the proposed methodology was applied to the hard science sectors of the Italian university research system for the period 2004-2006.
  6. Abramo, G.; D'Angelo, C.A.; Di Costa, F.: Testing the trade-off between productivity and quality in research activities (2009) 0.00
    0.0046599186 = product of:
      0.023299592 = sum of:
        0.023299592 = weight(_text_:system in 3317) [ClassicSimilarity], result of:
          0.023299592 = score(doc=3317,freq=2.0), product of:
            0.13391352 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.04251826 = queryNorm
            0.17398985 = fieldWeight in 3317, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3317)
      0.2 = coord(1/5)
    
    Abstract
    In recent years there has been an increasingly pressing need for the evaluation of results from public-sector research activity, particularly to permit the efficient allocation of ever scarcer resources. Many of the studies and evaluation exercises that have been conducted at the national and international levels emphasize the quality dimension of research output, while neglecting that of productivity. This work is intended to test for the possible existence of correlation between quantity and quality of scientific production and determine whether the most productive researchers are also those that achieve results that are qualitatively better than those of their colleagues. The analysis proposed refers to the entire Italian university system and is based on the observation of production in the hard sciences by more than 26,000 researchers in the period 2001-2005. The results show that the output of more-productive researchers is superior in quality than that of less-productive researchers. The relation between productivity and quality results is largely insensitive to the types of indicators or the test methods applied and also seems to differ little among the various disciplines examined.