Search (58 results, page 1 of 3)

  • × author_ss:"Thelwall, M."
  • × language_ss:"e"
  • × year_i:[2010 TO 2020}
  1. Li, X.; Thelwall, M.; Kousha, K.: ¬The role of arXiv, RePEc, SSRN and PMC in formal scholarly communication (2015) 0.06
    0.058023144 = product of:
      0.096705236 = sum of:
        0.020760437 = weight(_text_:of in 2593) [ClassicSimilarity], result of:
          0.020760437 = score(doc=2593,freq=20.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.27317715 = fieldWeight in 2593, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2593)
        0.059483737 = weight(_text_:subject in 2593) [ClassicSimilarity], result of:
          0.059483737 = score(doc=2593,freq=6.0), product of:
            0.17381717 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.04859849 = queryNorm
            0.34222013 = fieldWeight in 2593, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2593)
        0.016461061 = product of:
          0.032922123 = sum of:
            0.032922123 = weight(_text_:22 in 2593) [ClassicSimilarity], result of:
              0.032922123 = score(doc=2593,freq=2.0), product of:
                0.17018363 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04859849 = queryNorm
                0.19345059 = fieldWeight in 2593, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2593)
          0.5 = coord(1/2)
      0.6 = coord(3/5)
    
    Abstract
    Purpose The four major Subject Repositories (SRs), arXiv, Research Papers in Economics (RePEc), Social Science Research Network (SSRN) and PubMed Central (PMC), are all important within their disciplines but no previous study has systematically compared how often they are cited in academic publications. In response, the purpose of this paper is to report an analysis of citations to SRs from Scopus publications, 2000-2013. Design/methodology/approach Scopus searches were used to count the number of documents citing the four SRs in each year. A random sample of 384 documents citing the four SRs was then visited to investigate the nature of the citations. Findings Each SR was most cited within its own subject area but attracted substantial citations from other subject areas, suggesting that they are open to interdisciplinary uses. The proportion of documents citing each SR is continuing to increase rapidly, and the SRs all seem to attract substantial numbers of citations from more than one discipline. Research limitations/implications Scopus does not cover all publications, and most citations to documents found in the four SRs presumably cite the published version, when one exists, rather than the repository version. Practical implications SRs are continuing to grow and do not seem to be threatened by institutional repositories and so research managers should encourage their continued use within their core disciplines, including for research that aims at an audience in other disciplines. Originality/value This is the first simultaneous analysis of Scopus citations to the four most popular SRs.
    Date
    20. 1.2015 18:30:22
    Source
    Aslib journal of information management. 67(2015) no.6, S.614-635
  2. Thelwall, M.; Kousha, K.: ResearchGate: Disseminating, communicating, and measuring scholarship? (2015) 0.04
    0.042345606 = product of:
      0.10586401 = sum of:
        0.08656685 = weight(_text_:list in 1813) [ClassicSimilarity], result of:
          0.08656685 = score(doc=1813,freq=2.0), product of:
            0.25191793 = queryWeight, product of:
              5.183657 = idf(docFreq=673, maxDocs=44218)
              0.04859849 = queryNorm
            0.34363115 = fieldWeight in 1813, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.183657 = idf(docFreq=673, maxDocs=44218)
              0.046875 = fieldNorm(doc=1813)
        0.019297158 = weight(_text_:of in 1813) [ClassicSimilarity], result of:
          0.019297158 = score(doc=1813,freq=12.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.25392252 = fieldWeight in 1813, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=1813)
      0.4 = coord(2/5)
    
    Abstract
    ResearchGate is a social network site for academics to create their own profiles, list their publications, and interact with each other. Like Academia.edu, it provides a new way for scholars to disseminate their work and hence potentially changes the dynamics of informal scholarly communication. This article assesses whether ResearchGate usage and publication data broadly reflect existing academic hierarchies and whether individual countries are set to benefit or lose out from the site. The results show that rankings based on ResearchGate statistics correlate moderately well with other rankings of academic institutions, suggesting that ResearchGate use broadly reflects the traditional distribution of academic capital. Moreover, while Brazil, India, and some other countries seem to be disproportionately taking advantage of ResearchGate, academics in China, South Korea, and Russia may be missing opportunities to use ResearchGate to maximize the academic impact of their publications.
    Source
    Journal of the Association for Information Science and Technology. 66(2015) no.5, S.876-889
  3. Levitt, J.M.; Thelwall, M.; Oppenheim, C.: Variations between subjects in the extent to which the social sciences have become more interdisciplinary (2011) 0.04
    0.035778537 = product of:
      0.08944634 = sum of:
        0.020760437 = weight(_text_:of in 4465) [ClassicSimilarity], result of:
          0.020760437 = score(doc=4465,freq=20.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.27317715 = fieldWeight in 4465, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4465)
        0.068685904 = weight(_text_:subject in 4465) [ClassicSimilarity], result of:
          0.068685904 = score(doc=4465,freq=8.0), product of:
            0.17381717 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.04859849 = queryNorm
            0.39516178 = fieldWeight in 4465, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4465)
      0.4 = coord(2/5)
    
    Abstract
    Increasing interdisciplinarity has been a policy objective since the 1990s, promoted by many governments and funding agencies, but the question is: How deeply has this affected the social sciences? Although numerous articles have suggested that research has become more interdisciplinary, yet no study has compared the extent to which the interdisciplinarity of different social science subjects has changed. To address this gap, changes in the level of interdisciplinarity since 1980 are investigated for subjects with many articles in the Social Sciences Citation Index (SSCI), using the percentage of cross-disciplinary citing documents (PCDCD) to evaluate interdisciplinarity. For the 14 SSCI subjects investigated, the median level of interdisciplinarity, as measured using cross-disciplinary citations, declined from 1980 to 1990, but rose sharply between 1990 and 2000, confirming previous research. This increase was not fully matched by an increase in the percentage of articles that were assigned to more than one subject category. Nevertheless, although on average the social sciences have recently become more interdisciplinary, the extent of this change varies substantially from subject to subject. The SSCI subject with the largest increase in interdisciplinarity between 1990 and 2000 was Information Science & Library Science (IS&LS) but there is evidence that the level of interdisciplinarity of IS&LS increased less quickly during the first decade of this century.
    Source
    Journal of the American Society for Information Science and Technology. 62(2011) no.6, S.1118-1129
  4. Kousha, K.; Thelwall, M.: Can Amazon.com reviews help to assess the wider impacts of books? (2016) 0.03
    0.028275374 = product of:
      0.070688434 = sum of:
        0.029476898 = weight(_text_:of in 2768) [ClassicSimilarity], result of:
          0.029476898 = score(doc=2768,freq=28.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.38787308 = fieldWeight in 2768, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=2768)
        0.041211538 = weight(_text_:subject in 2768) [ClassicSimilarity], result of:
          0.041211538 = score(doc=2768,freq=2.0), product of:
            0.17381717 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.04859849 = queryNorm
            0.23709705 = fieldWeight in 2768, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.046875 = fieldNorm(doc=2768)
      0.4 = coord(2/5)
    
    Abstract
    Although citation counts are often used to evaluate the research impact of academic publications, they are problematic for books that aim for educational or cultural impact. To fill this gap, this article assesses whether a number of simple metrics derived from Amazon.com reviews of academic books could provide evidence of their impact. Based on a set of 2,739 academic monographs from 2008 and a set of 1,305 best-selling books in 15 Amazon.com academic subject categories, the existence of significant but low or moderate correlations between citations and numbers of reviews, combined with other evidence, suggests that online book reviews tend to reflect the wider popularity of a book rather than its academic impact, although there are substantial disciplinary differences. Metrics based on online reviews are therefore recommended for the evaluation of books that aim at a wide audience inside or outside academia when it is important to capture the broader impacts of educational or cultural activities and when they cannot be manipulated in advance of the evaluation.
    Source
    Journal of the Association for Information Science and Technology. 67(2016) no.3, S.566-581
  5. Kousha, K.; Thelwall, M.; Abdoli, M.: ¬The role of online videos in research communication : a content analysis of YouTube videos cited in academic publications (2012) 0.02
    0.021615213 = product of:
      0.054038033 = sum of:
        0.019695079 = weight(_text_:of in 382) [ClassicSimilarity], result of:
          0.019695079 = score(doc=382,freq=18.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.25915858 = fieldWeight in 382, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=382)
        0.034342952 = weight(_text_:subject in 382) [ClassicSimilarity], result of:
          0.034342952 = score(doc=382,freq=2.0), product of:
            0.17381717 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.04859849 = queryNorm
            0.19758089 = fieldWeight in 382, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.0390625 = fieldNorm(doc=382)
      0.4 = coord(2/5)
    
    Abstract
    Although there is some evidence that online videos are increasingly used by academics for informal scholarly communication and teaching, the extent to which they are used in published academic research is unknown. This article explores the extent to which YouTube videos are cited in academic publications and whether there are significant broad disciplinary differences in this practice. To investigate, we extracted the URL citations to YouTube videos from academic publications indexed by Scopus. A total of 1,808 Scopus publications cited at least one YouTube video, and there was a steady upward growth in citing online videos within scholarly publications from 2006 to 2011, with YouTube citations being most common within arts and humanities (0.3%) and the social sciences (0.2%). A content analysis of 551 YouTube videos cited by research articles indicated that in science (78%) and in medicine and health sciences (77%), over three fourths of the cited videos had either direct scientific (e.g., laboratory experiments) or scientific-related contents (e.g., academic lectures or education) whereas in the arts and humanities, about 80% of the YouTube videos had art, culture, or history themes, and in the social sciences, about 63% of the videos were related to news, politics, advertisements, and documentaries. This shows both the disciplinary differences and the wide variety of innovative research communication uses found for videos within the different subject areas.
    Source
    Journal of the American Society for Information Science and Technology. 63(2012) no.9, S.1710-1727
  6. Kousha, K.; Thelwall, M.; Rezaie, S.: Assessing the citation impact of books : the role of Google Books, Google Scholar, and Scopus (2011) 0.02
    0.020684952 = product of:
      0.05171238 = sum of:
        0.017369429 = weight(_text_:of in 4920) [ClassicSimilarity], result of:
          0.017369429 = score(doc=4920,freq=14.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.22855641 = fieldWeight in 4920, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4920)
        0.034342952 = weight(_text_:subject in 4920) [ClassicSimilarity], result of:
          0.034342952 = score(doc=4920,freq=2.0), product of:
            0.17381717 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.04859849 = queryNorm
            0.19758089 = fieldWeight in 4920, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4920)
      0.4 = coord(2/5)
    
    Abstract
    Citation indictors are increasingly used in some subject areas to support peer review in the evaluation of researchers and departments. Nevertheless, traditional journal-based citation indexes may be inadequate for the citation impact assessment of book-based disciplines. This article examines whether online citations from Google Books and Google Scholar can provide alternative sources of citation evidence. To investigate this, we compared the citation counts to 1,000 books submitted to the 2008 U.K. Research Assessment Exercise (RAE) from Google Books and Google Scholar with Scopus citations across seven book-based disciplines (archaeology; law; politics and international studies; philosophy; sociology; history; and communication, cultural, and media studies). Google Books and Google Scholar citations to books were 1.4 and 3.2 times more common than were Scopus citations, and their medians were more than twice and three times as high as were Scopus median citations, respectively. This large number of citations is evidence that in book-oriented disciplines in the social sciences, arts, and humanities, online book citations may be sufficiently numerous to support peer review for research evaluation, at least in the United Kingdom.
    Source
    Journal of the American Society for Information Science and Technology. 62(2011) no.11, S.2147-2164
  7. Kousha, K.; Thelwall, M.: ¬An automatic method for extracting citations from Google Books (2015) 0.02
    0.01960912 = product of:
      0.049022797 = sum of:
        0.014679846 = weight(_text_:of in 1658) [ClassicSimilarity], result of:
          0.014679846 = score(doc=1658,freq=10.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.19316542 = fieldWeight in 1658, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1658)
        0.034342952 = weight(_text_:subject in 1658) [ClassicSimilarity], result of:
          0.034342952 = score(doc=1658,freq=2.0), product of:
            0.17381717 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.04859849 = queryNorm
            0.19758089 = fieldWeight in 1658, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1658)
      0.4 = coord(2/5)
    
    Abstract
    Recent studies have shown that counting citations from books can help scholarly impact assessment and that Google Books (GB) is a useful source of such citation counts, despite its lack of a public citation index. Searching GB for citations produces approximate matches, however, and so its raw results need time-consuming human filtering. In response, this article introduces a method to automatically remove false and irrelevant matches from GB citation searches in addition to introducing refinements to a previous GB manual citation extraction method. The method was evaluated by manual checking of sampled GB results and comparing citations to about 14,500 monographs in the Thomson Reuters Book Citation Index (BKCI) against automatically extracted citations from GB across 24 subject areas. GB citations were 103% to 137% as numerous as BKCI citations in the humanities, except for tourism (72%) and linguistics (91%), 46% to 85% in social sciences, but only 8% to 53% in the sciences. In all cases, however, GB had substantially more citing books than did BKCI, with BKCI's results coming predominantly from journal articles. Moderate correlations between the GB and BKCI citation counts in social sciences and humanities, with most BKCI results coming from journal articles rather than books, suggests that they could measure the different aspects of impact, however.
    Source
    Journal of the Association for Information Science and Technology. 66(2015) no.2, S.309-320
  8. Thelwall, M.; Buckley, K.; Paltoglou, G.: Sentiment in Twitter events (2011) 0.02
    0.017866319 = product of:
      0.0446658 = sum of:
        0.024912525 = weight(_text_:of in 4345) [ClassicSimilarity], result of:
          0.024912525 = score(doc=4345,freq=20.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.32781258 = fieldWeight in 4345, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=4345)
        0.019753272 = product of:
          0.039506543 = sum of:
            0.039506543 = weight(_text_:22 in 4345) [ClassicSimilarity], result of:
              0.039506543 = score(doc=4345,freq=2.0), product of:
                0.17018363 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04859849 = queryNorm
                0.23214069 = fieldWeight in 4345, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4345)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    The microblogging site Twitter generates a constant stream of communication, some of which concerns events of general interest. An analysis of Twitter may, therefore, give insights into why particular events resonate with the population. This article reports a study of a month of English Twitter posts, assessing whether popular events are typically associated with increases in sentiment strength, as seems intuitively likely. Using the top 30 events, determined by a measure of relative increase in (general) term usage, the results give strong evidence that popular events are normally associated with increases in negative sentiment strength and some evidence that peaks of interest in events have stronger positive sentiment than the time before the peak. It seems that many positive events, such as the Oscars, are capable of generating increased negative sentiment in reaction to them. Nevertheless, the surprisingly small average change in sentiment associated with popular events (typically 1% and only 6% for Tiger Woods' confessions) is consistent with events affording posters opportunities to satisfy pre-existing personal goals more often than eliciting instinctive reactions.
    Date
    22. 1.2011 14:27:06
    Source
    Journal of the American Society for Information Science and Technology. 62(2011) no.2, S.406-418
  9. Thelwall, M.; Maflahi, N.: Guideline references and academic citations as evidence of the clinical value of health research (2016) 0.02
    0.017866319 = product of:
      0.0446658 = sum of:
        0.024912525 = weight(_text_:of in 2856) [ClassicSimilarity], result of:
          0.024912525 = score(doc=2856,freq=20.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.32781258 = fieldWeight in 2856, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=2856)
        0.019753272 = product of:
          0.039506543 = sum of:
            0.039506543 = weight(_text_:22 in 2856) [ClassicSimilarity], result of:
              0.039506543 = score(doc=2856,freq=2.0), product of:
                0.17018363 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04859849 = queryNorm
                0.23214069 = fieldWeight in 2856, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2856)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    This article introduces a new source of evidence of the value of medical-related research: citations from clinical guidelines. These give evidence that research findings have been used to inform the day-to-day practice of medical staff. To identify whether citations from guidelines can give different information from that of traditional citation counts, this article assesses the extent to which references in clinical guidelines tend to be highly cited in the academic literature and highly read in Mendeley. Using evidence from the United Kingdom, references associated with the UK's National Institute of Health and Clinical Excellence (NICE) guidelines tended to be substantially more cited than comparable articles, unless they had been published in the most recent 3 years. Citation counts also seemed to be stronger indicators than Mendeley readership altmetrics. Hence, although presence in guidelines may be particularly useful to highlight the contributions of recently published articles, for older articles citation counts may already be sufficient to recognize their contributions to health in society.
    Date
    19. 3.2016 12:22:00
    Source
    Journal of the Association for Information Science and Technology. 67(2016) no.4, S.960-966
  10. Didegah, F.; Thelwall, M.: Co-saved, co-tweeted, and co-cited networks (2018) 0.02
    0.017866319 = product of:
      0.0446658 = sum of:
        0.024912525 = weight(_text_:of in 4291) [ClassicSimilarity], result of:
          0.024912525 = score(doc=4291,freq=20.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.32781258 = fieldWeight in 4291, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=4291)
        0.019753272 = product of:
          0.039506543 = sum of:
            0.039506543 = weight(_text_:22 in 4291) [ClassicSimilarity], result of:
              0.039506543 = score(doc=4291,freq=2.0), product of:
                0.17018363 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04859849 = queryNorm
                0.23214069 = fieldWeight in 4291, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4291)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Counts of tweets and Mendeley user libraries have been proposed as altmetric alternatives to citation counts for the impact assessment of articles. Although both have been investigated to discover whether they correlate with article citations, it is not known whether users tend to tweet or save (in Mendeley) the same kinds of articles that they cite. In response, this article compares pairs of articles that are tweeted, saved to a Mendeley library, or cited by the same user, but possibly a different user for each source. The study analyzes 1,131,318 articles published in 2012, with minimum tweeted (10), saved to Mendeley (100), and cited (10) thresholds. The results show surprisingly minor overall overlaps between the three phenomena. The importance of journals for Twitter and the presence of many bots at different levels of activity suggest that this site has little value for impact altmetrics. The moderate differences between patterns of saving and citation suggest that Mendeley can be used for some types of impact assessments, but sensitivity is needed for underlying differences.
    Date
    28. 7.2018 10:00:22
    Source
    Journal of the Association for Information Science and Technology. 69(2018) no.8, S.959-973
  11. Thelwall, M.; Sud, P.; Wilkinson, D.: Link and co-inlink network diagrams with URL citations or title mentions (2012) 0.01
    0.0148885995 = product of:
      0.0372215 = sum of:
        0.020760437 = weight(_text_:of in 57) [ClassicSimilarity], result of:
          0.020760437 = score(doc=57,freq=20.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.27317715 = fieldWeight in 57, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=57)
        0.016461061 = product of:
          0.032922123 = sum of:
            0.032922123 = weight(_text_:22 in 57) [ClassicSimilarity], result of:
              0.032922123 = score(doc=57,freq=2.0), product of:
                0.17018363 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04859849 = queryNorm
                0.19345059 = fieldWeight in 57, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=57)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Webometric network analyses have been used to map the connectivity of groups of websites to identify clusters, important sites or overall structure. Such analyses have mainly been based upon hyperlink counts, the number of hyperlinks between a pair of websites, although some have used title mentions or URL citations instead. The ability to automatically gather hyperlink counts from Yahoo! ceased in April 2011 and the ability to manually gather such counts was due to cease by early 2012, creating a need for alternatives. This article assesses URL citations and title mentions as possible replacements for hyperlinks in both binary and weighted direct link and co-inlink network diagrams. It also assesses three different types of data for the network connections: hit count estimates, counts of matching URLs, and filtered counts of matching URLs. Results from analyses of U.S. library and information science departments and U.K. universities give evidence that metrics based upon URLs or titles can be appropriate replacements for metrics based upon hyperlinks for both binary and weighted networks, although filtered counts of matching URLs are necessary to give the best results for co-title mention and co-URL citation network diagrams.
    Date
    6. 4.2012 18:16:22
    Source
    Journal of the American Society for Information Science and Technology. 63(2012) no.4, S.805-816
  12. Thelwall, M.: Are Mendeley reader counts high enough for research evaluations when articles are published? (2017) 0.01
    0.0148885995 = product of:
      0.0372215 = sum of:
        0.020760437 = weight(_text_:of in 3806) [ClassicSimilarity], result of:
          0.020760437 = score(doc=3806,freq=20.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.27317715 = fieldWeight in 3806, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3806)
        0.016461061 = product of:
          0.032922123 = sum of:
            0.032922123 = weight(_text_:22 in 3806) [ClassicSimilarity], result of:
              0.032922123 = score(doc=3806,freq=2.0), product of:
                0.17018363 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04859849 = queryNorm
                0.19345059 = fieldWeight in 3806, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3806)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Purpose Mendeley reader counts have been proposed as early indicators for the impact of academic publications. The purpose of this paper is to assess whether there are enough Mendeley readers for research evaluation purposes during the month when an article is first published. Design/methodology/approach Average Mendeley reader counts were compared to the average Scopus citation counts for 104,520 articles from ten disciplines during the second half of 2016. Findings Articles attracted, on average, between 0.1 and 0.8 Mendeley readers per article in the month in which they first appeared in Scopus. This is about ten times more than the average Scopus citation count. Research limitations/implications Other disciplines may use Mendeley more or less than the ten investigated here. The results are dependent on Scopus's indexing practices, and Mendeley reader counts can be manipulated and have national and seniority biases. Practical implications Mendeley reader counts during the month of publication are more powerful than Scopus citations for comparing the average impacts of groups of documents but are not high enough to differentiate between the impacts of typical individual articles. Originality/value This is the first multi-disciplinary and systematic analysis of Mendeley reader counts from the publication month of an article.
    Date
    20. 1.2015 18:30:22
    Source
    Aslib journal of information management. 69(2017) no.2, S.174-183
  13. Thelwall, M.; Buckley, K.; Paltoglou, G.; Cai, D.; Kappas, A.: Sentiment strength detection in short informal text (2010) 0.01
    0.014462456 = product of:
      0.03615614 = sum of:
        0.019695079 = weight(_text_:of in 4200) [ClassicSimilarity], result of:
          0.019695079 = score(doc=4200,freq=18.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.25915858 = fieldWeight in 4200, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4200)
        0.016461061 = product of:
          0.032922123 = sum of:
            0.032922123 = weight(_text_:22 in 4200) [ClassicSimilarity], result of:
              0.032922123 = score(doc=4200,freq=2.0), product of:
                0.17018363 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04859849 = queryNorm
                0.19345059 = fieldWeight in 4200, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4200)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    A huge number of informal messages are posted every day in social network sites, blogs, and discussion forums. Emotions seem to be frequently important in these texts for expressing friendship, showing social support or as part of online arguments. Algorithms to identify sentiment and sentiment strength are needed to help understand the role of emotion in this informal communication and also to identify inappropriate or anomalous affective utterances, potentially associated with threatening behavior to the self or others. Nevertheless, existing sentiment detection algorithms tend to be commercially oriented, designed to identify opinions about products rather than user behaviors. This article partly fills this gap with a new algorithm, SentiStrength, to extract sentiment strength from informal English text, using new methods to exploit the de facto grammars and spelling styles of cyberspace. Applied to MySpace comments and with a lookup table of term sentiment strengths optimized by machine learning, SentiStrength is able to predict positive emotion with 60.6% accuracy and negative emotion with 72.8% accuracy, both based upon strength scales of 1-5. The former, but not the latter, is better than baseline and a wide range of general machine learning approaches.
    Date
    22. 1.2011 14:29:23
    Footnote
    Vgl. auch das Erratum in: Journal of the American Society for Information Science and Technology. 62(2011) no.2, S.419
    Source
    Journal of the American Society for Information Science and Technology. 61(2010) no.12, S.2544-2558
  14. Thelwall, M.; Sud, P.: Mendeley readership counts : an investigation of temporal and disciplinary differences (2016) 0.01
    0.014203735 = product of:
      0.035509337 = sum of:
        0.015756065 = weight(_text_:of in 3211) [ClassicSimilarity], result of:
          0.015756065 = score(doc=3211,freq=8.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.20732689 = fieldWeight in 3211, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=3211)
        0.019753272 = product of:
          0.039506543 = sum of:
            0.039506543 = weight(_text_:22 in 3211) [ClassicSimilarity], result of:
              0.039506543 = score(doc=3211,freq=2.0), product of:
                0.17018363 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04859849 = queryNorm
                0.23214069 = fieldWeight in 3211, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3211)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Scientists and managers using citation-based indicators to help evaluate research cannot evaluate recent articles because of the time needed for citations to accrue. Reading occurs before citing, however, and so it makes sense to count readers rather than citations for recent publications. To assess this, Mendeley readers and citations were obtained for articles from 2004 to late 2014 in five broad categories (agriculture, business, decision science, pharmacy, and the social sciences) and 50 subcategories. In these areas, citation counts tended to increase with every extra year since publication, and readership counts tended to increase faster initially but then stabilize after about 5 years. The correlation between citations and readers was also higher for longer time periods, stabilizing after about 5 years. Although there were substantial differences between broad fields and smaller differences between subfields, the results confirm the value of Mendeley reader counts as early scientific impact indicators.
    Date
    16.11.2016 11:07:22
    Source
    Journal of the Association for Information Science and Technology. 67(2016) no.12, S.3036-3050
  15. Didegah, F.; Thelwall, M.: Determinants of research citation impact in nanoscience and nanotechnology (2013) 0.01
    0.00568093 = product of:
      0.02840465 = sum of:
        0.02840465 = weight(_text_:of in 737) [ClassicSimilarity], result of:
          0.02840465 = score(doc=737,freq=26.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.37376386 = fieldWeight in 737, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=737)
      0.2 = coord(1/5)
    
    Abstract
    This study investigates a range of metrics available when a nanoscience and nanotechnology article is published to see which metrics correlate more with the number of citations to the article. It also introduces the degree of internationality of journals and references as new metrics for this purpose. The journal impact factor; the impact of references; the internationality of authors, journals, and references; and the number of authors, institutions, and references were all calculated for papers published in nanoscience and nanotechnology journals in the Web of Science from 2007 to 2009. Using a zero-inflated negative binomial regression model on the data set, the impact factor of the publishing journal and the citation impact of the cited references were found to be the most effective determinants of citation counts in all four time periods. In the entire 2007 to 2009 period, apart from journal internationality and author numbers and internationality, all other predictor variables had significant effects on citation counts.
    Source
    Journal of the American Society for Information Science and Technology. 64(2013) no.5, S.1055-1064
  16. Larivière, V.; Sugimoto, C.R.; Macaluso, B.; Milojevi´c, S.; Cronin, B.; Thelwall, M.: arXiv E-prints and the journal of record : an analysis of roles and relationships (2014) 0.01
    0.005252022 = product of:
      0.026260108 = sum of:
        0.026260108 = weight(_text_:of in 1285) [ClassicSimilarity], result of:
          0.026260108 = score(doc=1285,freq=32.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.34554482 = fieldWeight in 1285, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1285)
      0.2 = coord(1/5)
    
    Abstract
    Since its creation in 1991, arXiv has become central to the diffusion of research in a number of fields. Combining data from the entirety of arXiv and the Web of Science (WoS), this article investigates (a) the proportion of papers across all disciplines that are on arXiv and the proportion of arXiv papers that are in the WoS, (b) the elapsed time between arXiv submission and journal publication, and (c) the aging characteristics and scientific impact of arXiv e-prints and their published version. It shows that the proportion of WoS papers found on arXiv varies across the specialties of physics and mathematics, and that only a few specialties make extensive use of the repository. Elapsed time between arXiv submission and journal publication has shortened but remains longer in mathematics than in physics. In physics, mathematics, as well as in astronomy and astrophysics, arXiv versions are cited more promptly and decay faster than WoS papers. The arXiv versions of papers-both published and unpublished-have lower citation rates than published papers, although there is almost no difference in the impact of the arXiv versions of published and unpublished papers.
    Source
    Journal of the Association for Information Science and Technology. 65(2014) no.6, S.1157-1169
  17. Thelwall, M.; Klitkou, A.; Verbeek, A.; Stuart, D.; Vincent, C.: Policy-relevant Webometrics for individual scientific fields (2010) 0.01
    0.0052256957 = product of:
      0.026128478 = sum of:
        0.026128478 = weight(_text_:of in 3574) [ClassicSimilarity], result of:
          0.026128478 = score(doc=3574,freq=22.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.34381276 = fieldWeight in 3574, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=3574)
      0.2 = coord(1/5)
    
    Abstract
    Despite over 10 years of research there is no agreement on the most suitable roles for Webometric indicators in support of research policy and almost no field-based Webometrics. This article partly fills these gaps by analyzing the potential of policy-relevant Webometrics for individual scientific fields with the help of 4 case studies. Although Webometrics cannot provide robust indicators of knowledge flows or research impact, it can provide some evidence of networking and mutual awareness. The scope of Webometrics is also relatively wide, including not only research organizations and firms but also intermediary groups like professional associations, Web portals, and government agencies. Webometrics can, therefore, provide evidence about the research process to compliment peer review, bibliometric, and patent indicators: tracking the early, mainly prepublication development of new fields and research funding initiatives, assessing the role and impact of intermediary organizations and the need for new ones, and monitoring the extent of mutual awareness in particular research areas.
    Source
    Journal of the American Society for Information Science and Technology. 61(2010) no.7, S.1464-1475
  18. Thelwall, M.; Goriunova, O.; Vis, F.; Faulkner, S.; Burns, A.; Aulich, J.; Mas-Bleda, A.; Stuart, E.; D'Orazio, F.: Chatting through pictures : a classification of images tweeted in one week in the UK and USA (2016) 0.01
    0.005085248 = product of:
      0.02542624 = sum of:
        0.02542624 = weight(_text_:of in 3215) [ClassicSimilarity], result of:
          0.02542624 = score(doc=3215,freq=30.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.33457235 = fieldWeight in 3215, product of:
              5.477226 = tf(freq=30.0), with freq of:
                30.0 = termFreq=30.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3215)
      0.2 = coord(1/5)
    
    Abstract
    Twitter is used by a substantial minority of the populations of many countries to share short messages, sometimes including images. Nevertheless, despite some research into specific images, such as selfies, and a few news stories about specific tweeted photographs, little is known about the types of images that are routinely shared. In response, this article reports a content analysis of random samples of 800 images tweeted from the UK or USA during a week at the end of 2014. Although most images were photographs, a substantial minority were hybrid or layered image forms: phone screenshots, collages, captioned pictures, and pictures of text messages. About half were primarily of one or more people, including 10% that were selfies, but a wide variety of other things were also pictured. Some of the images were for advertising or to share a joke but in most cases the purpose of the tweet seemed to be to share the minutiae of daily lives, performing the function of chat or gossip, sometimes in innovative ways.
    Source
    Journal of the Association for Information Science and Technology. 67(2016) no.11, S.2575-2586
  19. Kousha, K.; Thelwall, M.: ¬An automatic method for assessing the teaching impact of books from online academic syllabi (2016) 0.01
    0.005085248 = product of:
      0.02542624 = sum of:
        0.02542624 = weight(_text_:of in 3226) [ClassicSimilarity], result of:
          0.02542624 = score(doc=3226,freq=30.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.33457235 = fieldWeight in 3226, product of:
              5.477226 = tf(freq=30.0), with freq of:
                30.0 = termFreq=30.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3226)
      0.2 = coord(1/5)
    
    Abstract
    Scholars writing books that are widely used to support teaching in higher education may be undervalued because of a lack of evidence of teaching value. Although sales data may give credible evidence for textbooks, these data may poorly reflect educational uses of other types of books. As an alternative, this article proposes a method to search automatically for mentions of books in online academic course syllabi based on Bing searches for syllabi mentioning a given book, filtering out false matches through an extensive set of rules. The method had an accuracy of over 90% based on manual checks of a sample of 2,600 results from the initial Bing searches. Over one third of about 14,000 monographs checked had one or more academic syllabus mention, with more in the arts and humanities (56%) and social sciences (52%). Low but significant correlations between syllabus mentions and citations across most fields, except the social sciences, suggest that books tend to have different levels of impact for teaching and research. In conclusion, the automatic syllabus search method gives a new way to estimate the educational utility of books in a way that sales data and citation counts cannot.
    Source
    Journal of the Association for Information Science and Technology. 67(2016) no.12, S.2993-3007
  20. Kousha, K.; Thelwall, M.: Are wikipedia citations important evidence of the impact of scholarly articles and books? (2017) 0.00
    0.0049128165 = product of:
      0.024564082 = sum of:
        0.024564082 = weight(_text_:of in 3440) [ClassicSimilarity], result of:
          0.024564082 = score(doc=3440,freq=28.0), product of:
            0.07599624 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.04859849 = queryNorm
            0.32322758 = fieldWeight in 3440, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3440)
      0.2 = coord(1/5)
    
    Abstract
    Individual academics and research evaluators often need to assess the value of published research. Although citation counts are a recognized indicator of scholarly impact, alternative data is needed to provide evidence of other types of impact, including within education and wider society. Wikipedia is a logical choice for both of these because the role of a general encyclopaedia is to be an understandable repository of facts about a diverse array of topics and hence it may cite research to support its claims. To test whether Wikipedia could provide new evidence about the impact of scholarly research, this article counted citations to 302,328 articles and 18,735 monographs in English indexed by Scopus in the period 2005 to 2012. The results show that citations from Wikipedia to articles are too rare for most research evaluation purposes, with only 5% of articles being cited in all fields. In contrast, a third of monographs have at least one citation from Wikipedia, with the most in the arts and humanities. Hence, Wikipedia citations can provide extra impact evidence for academic monographs. Nevertheless, the results may be relatively easily manipulated and so Wikipedia is not recommended for evaluations affecting stakeholder interests.
    Source
    Journal of the Association for Information Science and Technology. 68(2017) no.3, S.762-779