Search (59 results, page 2 of 3)

  • × author_ss:"Egghe, L."
  • × language_ss:"e"
  1. Egghe, L.; Rousseau, R.: ¬A measure for the cohesion of weighted networks (2003) 0.00
    0.002277429 = product of:
      0.02049686 = sum of:
        0.02049686 = weight(_text_:of in 5157) [ClassicSimilarity], result of:
          0.02049686 = score(doc=5157,freq=30.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.33457235 = fieldWeight in 5157, product of:
              5.477226 = tf(freq=30.0), with freq of:
                30.0 = termFreq=30.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5157)
      0.11111111 = coord(1/9)
    
    Abstract
    Measurement of the degree of interconnectedness in graph like networks of hyperlinks or citations can indicate the existence of research fields and assist in comparative evaluation of research efforts. In this issue we begin with Egghe and Rousseau who review compactness measures and investigate the compactness of a network as a weighted graph with dissimilarity values characterizing the arcs between nodes. They make use of a generalization of the Botofogo, Rivlin, Shneiderman, (BRS) compaction measure which treats the distance between unreachable nodes not as infinity but rather as the number of nodes in the network. The dissimilarity values are determined by summing the reciprocals of the weights of the arcs in the shortest chain between two nodes where no weight is smaller than one. The BRS measure is then the maximum value for the sum of the dissimilarity measures less the actual sum divided by the difference between the maximum and minimum. The Wiener index, the sum of all elements in the dissimilarity matrix divided by two, is then computed for Small's particle physics co-citation data as well as the BRS measure, the dissimilarity values and shortest paths. The compactness measure for the weighted network is smaller than for the un-weighted. When the bibliographic coupling network is utilized it is shown to be less compact than the co-citation network which indicates that the new measure produces results that confirm to an obvious case.
    Source
    Journal of the American Society for Information Science and technology. 54(2003) no.3, S.193-202
  2. Egghe, L.; Ravichandra Rao, I.K.: Duality revisited : construction of fractional frequency distributions based on two dual Lotka laws (2002) 0.00
    0.0022314154 = product of:
      0.020082738 = sum of:
        0.020082738 = weight(_text_:of in 1006) [ClassicSimilarity], result of:
          0.020082738 = score(doc=1006,freq=20.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.32781258 = fieldWeight in 1006, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=1006)
      0.11111111 = coord(1/9)
    
    Abstract
    Fractional frequency distributions of, for example, authors with a certain (fractional) number of papers are very irregular and, therefore, not easy to model or to explain. This article gives a first attempt to this by assuming two simple Lotka laws (with exponent 2): one for the number of authors with n papers (total count here) and one for the number of papers with n authors, n E N. Based an an earlier made convolution model of Egghe, interpreted and reworked now for discrete scores, we are able to produce theoretical fractional frequency distributions with only one parameter, which are in very close agreement with the practical ones as found in a large dataset produced earlier by Rao. The article also shows that (irregular) fractional frequency distributions are a consequence of Lotka's law, and are not examples of breakdowns of this famous historical law.
    Source
    Journal of the American Society for Information Science and technology. 53(2002) no.10, S.789-801
  3. Egghe, L.: Mathematical theory of the h- and g-index in case of fractional counting of authorship (2008) 0.00
    0.0022314154 = product of:
      0.020082738 = sum of:
        0.020082738 = weight(_text_:of in 2004) [ClassicSimilarity], result of:
          0.020082738 = score(doc=2004,freq=20.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.32781258 = fieldWeight in 2004, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=2004)
      0.11111111 = coord(1/9)
    
    Abstract
    This article studies the h-index (Hirsch index) and the g-index of authors, in case one counts authorship of the cited articles in a fractional way. There are two ways to do this: One counts the citations to these papers in a fractional way or one counts the ranks of the papers in a fractional way as credit for an author. In both cases, we define the fractional h- and g-indexes, and we present inequalities (both upper and lower bounds) between these fractional h- and g-indexes and their corresponding unweighted values (also involving, of course, the coauthorship distribution). Wherever applicable, examples and counterexamples are provided. In a concrete example (the publication citation list of the present author), we make explicit calculations of these fractional h- and g-indexes and show that they are not very different from the unweighted ones.
    Source
    Journal of the American Society for Information Science and Technology. 59(2008) no.10, S.1608-1616
  4. Egghe, L.: Properties of the n-overlap vector and n-overlap similarity theory (2006) 0.00
    0.0022002053 = product of:
      0.019801848 = sum of:
        0.019801848 = weight(_text_:of in 194) [ClassicSimilarity], result of:
          0.019801848 = score(doc=194,freq=28.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.32322758 = fieldWeight in 194, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=194)
      0.11111111 = coord(1/9)
    
    Abstract
    In the first part of this article the author defines the n-overlap vector whose coordinates consist of the fraction of the objects (e.g., books, N-grams, etc.) that belong to 1, 2, , n sets (more generally: families) (e.g., libraries, databases, etc.). With the aid of the Lorenz concentration theory, a theory of n-overlap similarity is conceived together with corresponding measures, such as the generalized Jaccard index (generalizing the well-known Jaccard index in case n 5 2). Next, the distributional form of the n-overlap vector is determined assuming certain distributions of the object's and of the set (family) sizes. In this section the decreasing power law and decreasing exponential distribution is explained for the n-overlap vector. Both item (token) n-overlap and source (type) n-overlap are studied. The n-overlap properties of objects indexed by a hierarchical system (e.g., books indexed by numbers from a UDC or Dewey system or by N-grams) are presented in the final section. The author shows how the results given in the previous section can be applied as well as how the Lorenz order of the n-overlap vector is respected by an increase or a decrease of the level of refinement in the hierarchical system (e.g., the value N in N-grams).
    Source
    Journal of the American Society for Information Science and Technology. 57(2006) no.9, S.1165-1177
  5. Egghe, L.: ¬The amount of actions needed for shelving and reshelving (1996) 0.00
    0.0021037988 = product of:
      0.018934188 = sum of:
        0.018934188 = weight(_text_:of in 4394) [ClassicSimilarity], result of:
          0.018934188 = score(doc=4394,freq=10.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.3090647 = fieldWeight in 4394, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=4394)
      0.11111111 = coord(1/9)
    
    Abstract
    Discusses the number of actions (or time) needed to organize library shelves. Studies 2 types pf problem: organizing a library shelf out of an unordered pile of books, and putting an existing shelf of books in the rough order. Uses results from information theory as well as from rank order statistics (runs). Draws conclusions about the advised frequency with which these actions should be undertaken
  6. Egghe, L.: Informetric explanation of some Leiden Ranking graphs (2014) 0.00
    0.0021037988 = product of:
      0.018934188 = sum of:
        0.018934188 = weight(_text_:of in 1236) [ClassicSimilarity], result of:
          0.018934188 = score(doc=1236,freq=10.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.3090647 = fieldWeight in 1236, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=1236)
      0.11111111 = coord(1/9)
    
    Abstract
    The S-shaped functional relation between the mean citation score and the proportion of top 10% publications for the 500 Leiden Ranking universities is explained using results of the shifted Lotka function. Also the concave or convex relation between the proportion of top 100?% publications, for different fractions ?, is explained using the obtained new informetric model.
    Source
    Journal of the Association for Information Science and Technology. 65(2014) no.4, S.737-741
  7. Egghe, L.: ¬The influence of transformations on the h-index and the g-index (2008) 0.00
    0.0020165213 = product of:
      0.018148692 = sum of:
        0.018148692 = weight(_text_:of in 1881) [ClassicSimilarity], result of:
          0.018148692 = score(doc=1881,freq=12.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.29624295 = fieldWeight in 1881, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1881)
      0.11111111 = coord(1/9)
    
    Abstract
    In a previous article, we introduced a general transformation on sources and one on items in an arbitrary information production process (IPP). In this article, we investigate the influence of these transformations on the h-index and on the g-index. General formulae that describe this influence are presented. These are applied to the case that the size-frequency function is Lotkaian (i.e., is a decreasing power function). We further show that the h-index of the transformed IPP belongs to the interval bounded by the two transformations of the h-index of the original IPP, and we also show that this property is not true for the g-index.
    Source
    Journal of the American Society for Information Science and Technology. 59(2008) no.8, S.1304-1312
  8. Egghe, L.; Liang, L.; Rousseau, R.: Fundamental properties of rhythm sequences (2008) 0.00
    0.0020165213 = product of:
      0.018148692 = sum of:
        0.018148692 = weight(_text_:of in 1965) [ClassicSimilarity], result of:
          0.018148692 = score(doc=1965,freq=12.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.29624295 = fieldWeight in 1965, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1965)
      0.11111111 = coord(1/9)
    
    Abstract
    Fundamental mathematical properties of rhythm sequences are studied. In particular, a set of three axioms for valid rhythm indicators is proposed, and it is shown that the R-indicator satisfies only two out of three but that the R-indicator satisfies all three. This fills a critical, logical gap in the study of these indicator sequences. Matrices leading to a constant R-sequence are called baseline matrices. They are characterized as matrices with constant w-year diachronous impact factors. The relation with classical impact factors is clarified. Using regression analysis matrices with a rhythm sequence that is on average equal to 1 (smaller than 1, larger than 1) are characterized.
    Source
    Journal of the American Society for Information Science and Technology. 59(2008) no.9, S.1469-1478
  9. Egghe, L.: ¬A new short proof of Naranan's theorem, explaining Lotka's law and Zipf's law (2010) 0.00
    0.0020165213 = product of:
      0.018148692 = sum of:
        0.018148692 = weight(_text_:of in 3432) [ClassicSimilarity], result of:
          0.018148692 = score(doc=3432,freq=12.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.29624295 = fieldWeight in 3432, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3432)
      0.11111111 = coord(1/9)
    
    Abstract
    Naranan's important theorem, published in Nature in 1970, states that if the number of journals grows exponentially and if the number of articles in each journal grows exponentially (at the same rate for each journal), then the system satisfies Lotka's law and a formula for the Lotka's exponent is given in function of the growth rates of the journals and the articles. This brief communication re-proves this result by showing that the system satisfies Zipf's law, which is equivalent with Lotka's law. The proof is short and algebraic and does not use infinitesimal arguments.
    Source
    Journal of the American Society for Information Science and Technology. 61(2010) no.12, S.2581-2583
  10. Egghe, L.: Remarks on the paper by A. De Visscher, "what does the g-index really measure?" (2012) 0.00
    0.0020165213 = product of:
      0.018148692 = sum of:
        0.018148692 = weight(_text_:of in 463) [ClassicSimilarity], result of:
          0.018148692 = score(doc=463,freq=12.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.29624295 = fieldWeight in 463, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=463)
      0.11111111 = coord(1/9)
    
    Abstract
    The author presents a different view on properties of impact measures than given in the paper of De Visscher (2011). He argues that a good impact measure works better when citations are concentrated rather than spread out over articles. The author also presents theoretical evidence that the g-index and the R-index can be close to the square root of the total number of citations, whereas this is not the case for the A-index. Here the author confirms an assertion of De Visscher.
    Source
    Journal of the American Society for Information Science and Technology. 63(2012) no.10, S.2118-2121
  11. Egghe, L.: Expansion of the field of informetrics : the second special issue (2006) 0.00
    0.0019958385 = product of:
      0.017962547 = sum of:
        0.017962547 = weight(_text_:of in 7119) [ClassicSimilarity], result of:
          0.017962547 = score(doc=7119,freq=4.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.2932045 = fieldWeight in 7119, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.09375 = fieldNorm(doc=7119)
      0.11111111 = coord(1/9)
    
  12. Egghe, L.: Expansion of the field of informetrics : origins and consequences (2005) 0.00
    0.0019958385 = product of:
      0.017962547 = sum of:
        0.017962547 = weight(_text_:of in 1910) [ClassicSimilarity], result of:
          0.017962547 = score(doc=1910,freq=4.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.2932045 = fieldWeight in 1910, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.09375 = fieldNorm(doc=1910)
      0.11111111 = coord(1/9)
    
  13. Egghe, L.: Relations between the continuous and the discrete Lotka power function (2005) 0.00
    0.0019958385 = product of:
      0.017962547 = sum of:
        0.017962547 = weight(_text_:of in 3464) [ClassicSimilarity], result of:
          0.017962547 = score(doc=3464,freq=16.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.2932045 = fieldWeight in 3464, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=3464)
      0.11111111 = coord(1/9)
    
    Abstract
    The discrete Lotka power function describes the number of sources (e.g., authors) with n = 1, 2, 3, ... items (e.g., publications). As in econometrics, informetrics theory requires functions of a continuous variable j, replacing the discrete variable n. Now j represents item densities instead of number of items. The continuous Lotka power function describes the density of sources with item density j. The discrete Lotka function one obtains from data, obtained empirically; the continuous Lotka function is the one needed when one wants to apply Lotkaian informetrics, i.e., to determine properties that can be derived from the (continuous) model. It is, hence, important to know the relations between the two models. We show that the exponents of the discrete Lotka function (if not too high, i.e., within limits encountered in practice) and of the continuous Lotka function are approximately the same. This is important to know in applying theoretical results (from the continuous model), derived from practical data.
    Source
    Journal of the American Society for Information Science and Technology. 56(2005) no.7, S.664-668
  14. Egghe, L.; Leydesdorff, L.: ¬The relation between Pearson's correlation coefficient r and Salton's cosine measure (2009) 0.00
    0.0019958385 = product of:
      0.017962547 = sum of:
        0.017962547 = weight(_text_:of in 2803) [ClassicSimilarity], result of:
          0.017962547 = score(doc=2803,freq=16.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.2932045 = fieldWeight in 2803, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=2803)
      0.11111111 = coord(1/9)
    
    Abstract
    The relation between Pearson's correlation coefficient and Salton's cosine measure is revealed based on the different possible values of the division of the L1-norm and the L2-norm of a vector. These different values yield a sheaf of increasingly straight lines which together form a cloud of points, being the investigated relation. The theoretical results are tested against the author co-citation relations among 24 informetricians for whom two matrices can be constructed, based on co-citations: the asymmetric occurrence matrix and the symmetric co-citation matrix. Both examples completely confirm the theoretical results. The results enable us to specify an algorithm that provides a threshold value for the cosine above which none of the corresponding Pearson correlations would be negative. Using this threshold value can be expected to optimize the visualization of the vector space.
    Source
    Journal of the American Society for Information Science and Technology. 60(2009) no.5, S.1027-1036
  15. Egghe, L.; Guns, R.; Rousseau, R.: Thoughts on uncitedness : Nobel laureates and Fields medalists as case studies (2011) 0.00
    0.0019958385 = product of:
      0.017962547 = sum of:
        0.017962547 = weight(_text_:of in 4994) [ClassicSimilarity], result of:
          0.017962547 = score(doc=4994,freq=16.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.2932045 = fieldWeight in 4994, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=4994)
      0.11111111 = coord(1/9)
    
    Abstract
    Contrary to what one might expect, Nobel laureates and Fields medalists have a rather large fraction (10% or more) of uncited publications. This is the case for (in total) 75 examined researchers from the fields of mathematics (Fields medalists), physics, chemistry, and physiology or medicine (Nobel laureates). We study several indicators for these researchers, including the h-index, total number of publications, average number of citations per publication, the number (and fraction) of uncited publications, and their interrelations. The most remarkable result is a positive correlation between the h-index and the number of uncited articles. We also present a Lotkaian model, which partially explains the empirically found regularities.
    Footnote
    Vgl.: Erratum. In: Journal of the American Society for Information Science and Technology. 63(2012) no.2, S.429.
    Source
    Journal of the American Society for Information Science and Technology. 62(2011) no.8, S.1637-1644
  16. Egghe, L.: Note on a possible decomposition of the h-Index (2013) 0.00
    0.0019958385 = product of:
      0.017962547 = sum of:
        0.017962547 = weight(_text_:of in 683) [ClassicSimilarity], result of:
          0.017962547 = score(doc=683,freq=4.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.2932045 = fieldWeight in 683, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.09375 = fieldNorm(doc=683)
      0.11111111 = coord(1/9)
    
    Source
    Journal of the American Society for Information Science and Technology. 64(2013) no.4, S.871
  17. Egghe, L.: Type/Token-Taken informetrics (2003) 0.00
    0.0019502735 = product of:
      0.017552461 = sum of:
        0.017552461 = weight(_text_:of in 1608) [ClassicSimilarity], result of:
          0.017552461 = score(doc=1608,freq=22.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.28651062 = fieldWeight in 1608, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1608)
      0.11111111 = coord(1/9)
    
    Abstract
    Type/Token-Taken informetrics is a new part of informetrics that studies the use of items rather than the items itself. Here, items are the objects that are produced by the sources (e.g., journals producing articles, authors producing papers, etc.). In linguistics a source is also called a type (e.g., a word), and an item a token (e.g., the use of words in texts). In informetrics, types that occur often, for example, in a database will also be requested often, for example, in information retrieval. The relative use of these occurrences will be higher than their relative occurrences itself; hence, the name Type/ Token-Taken informetrics. This article studies the frequency distribution of Type/Token-Taken informetrics, starting from the one of Type/Token informetrics (i.e., source-item relationships). We are also studying the average number my* of item uses in Type/Token-Taken informetrics and compare this with the classical average number my in Type/Token informetrics. We show that my* >= my always, and that my* is an increasing function of my. A method is presented to actually calculate my* from my, and a given a, which is the exponent in Lotka's frequency distribution of Type/Token informetrics. We leave open the problem of developing non-Lotkaian Type/TokenTaken informetrics.
    Source
    Journal of the American Society for Information Science and technology. 54(2003) no.7, S.603-610
  18. Egghe, L.; Rousseau, R.: ¬The influence of publication delays on the observed aging distribution of scientific literature (2000) 0.00
    0.0018816947 = product of:
      0.016935252 = sum of:
        0.016935252 = weight(_text_:of in 4385) [ClassicSimilarity], result of:
          0.016935252 = score(doc=4385,freq=8.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.27643585 = fieldWeight in 4385, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=4385)
      0.11111111 = coord(1/9)
    
    Abstract
    Observed aging curves are influenced by publication delays. In this article, we show how the 'undisturbed' aging function and the publication delay combine to give the observed aging function. This combination is performed by a mathematical operation known as convolution. Examples are given, such as the convolution of 2 Poisson distributions, 2 exponential distributions, a 2 lognormal distributions. A paradox is observed between theory and real data
    Source
    Journal of the American Society for Information Science. 51(2000) no.2, S.158-165
  19. Egghe, L.: Empirical and combinatorial study of country occurrences in multi-authored papers (2006) 0.00
    0.0018816947 = product of:
      0.016935252 = sum of:
        0.016935252 = weight(_text_:of in 81) [ClassicSimilarity], result of:
          0.016935252 = score(doc=81,freq=32.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.27643585 = fieldWeight in 81, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=81)
      0.11111111 = coord(1/9)
    
    Abstract
    Papers written by several authors can be classified according to the countries of the author affiliations. The empirical part of this paper consists of two datasets. One dataset consists of 1,035 papers retrieved via the search "pedagog*" in the years 2004 and 2005 (up to October) in Academic Search Elite which is a case where phi(m) = the number of papers with m =1, 2,3 ... authors is decreasing, hence most of the papers have a low number of authors. Here we find that #, m = the number of times a country occurs j times in a m-authored paper, j =1, ..., m-1 is decreasing and that # m, m is much higher than all the other #j, m values. The other dataset consists of 3,271 papers retrieved via the search "enzyme" in the year 2005 (up to October) in the same database which is a case of a non-decreasing phi(m): most papers have 3 or 4 authors and we even find many papers with a much higher number of authors. In this case we show again that # m, m is much higher than the other #j, m values but that #j, m is not decreasing anymore in j =1, ..., m-1, although #1, m is (apart from # m, m) the largest number amongst the #j,m. The combinatorial part gives a proof of the fact that #j,m decreases for j = 1, m-1, supposing that all cases are equally possible. This shows that the first dataset is more conform with this model than the second dataset. Explanations for these findings are given. From the data we also find the (we think: new) distribution of number of papers with n =1, 2,3,... countries (i.e. where there are n different countries involved amongst the m (a n) authors of a paper): a fast decreasing function e.g. as a power law with a very large Lotka exponent.
  20. Egghe, L.: ¬A noninformetric analysis of the relationship between citation age and journal productivity (2001) 0.00
    0.0018669361 = product of:
      0.016802425 = sum of:
        0.016802425 = weight(_text_:of in 5685) [ClassicSimilarity], result of:
          0.016802425 = score(doc=5685,freq=14.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.2742677 = fieldWeight in 5685, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=5685)
      0.11111111 = coord(1/9)
    
    Abstract
    A problem, raised by Wallace (JASIS, 37,136-145,1986), on the relation between the journal's median citation age and its number of articles is studied. Leaving open the problem as such, we give a statistical explanation of this relationship, when replacing "median" by "mean" in Wallace's problem. The cloud of points, found by Wallace, is explained in this sense that the points are scattered over the area in first quadrant, limited by a curve of the form y=1 + E/x**2 where E is a constant. This curve is obtained by using the Central Limit Theorem in statistics and, hence, has no intrinsic informetric foundation. The article closes with some reflections on explanations of regularities in informetrics, based on statistical, probabilistic or informetric results, or on a combination thereof
    Source
    Journal of the American Society for Information Science and technology. 52(2001) no.5, S.371-377