Search (9 results, page 1 of 1)

  • × author_ss:"Vaughan, L."
  1. Vaughan, L.; Thelwall, M.: ¬A modelling approach to uncover hyperlink patterns : the case of Canadian universities (2005) 0.03
    0.028032223 = product of:
      0.09811278 = sum of:
        0.036320645 = weight(_text_:management in 1014) [ClassicSimilarity], result of:
          0.036320645 = score(doc=1014,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.2606825 = fieldWeight in 1014, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1014)
        0.061792135 = weight(_text_:case in 1014) [ClassicSimilarity], result of:
          0.061792135 = score(doc=1014,freq=2.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.34001783 = fieldWeight in 1014, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1014)
      0.2857143 = coord(2/7)
    
    Source
    Information processing and management. 41(2005) no.2, S.347-360
  2. Vaughan, L.; Chen, Y.: Data mining from web search queries : a comparison of Google trends and Baidu index (2015) 0.01
    0.00919453 = product of:
      0.06436171 = sum of:
        0.06436171 = sum of:
          0.03635913 = weight(_text_:studies in 1605) [ClassicSimilarity], result of:
            0.03635913 = score(doc=1605,freq=2.0), product of:
              0.16494368 = queryWeight, product of:
                3.9902744 = idf(docFreq=2222, maxDocs=44218)
                0.041336425 = queryNorm
              0.22043361 = fieldWeight in 1605, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.9902744 = idf(docFreq=2222, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1605)
          0.028002575 = weight(_text_:22 in 1605) [ClassicSimilarity], result of:
            0.028002575 = score(doc=1605,freq=2.0), product of:
              0.14475311 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.041336425 = queryNorm
              0.19345059 = fieldWeight in 1605, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1605)
      0.14285715 = coord(1/7)
    
    Abstract
    Numerous studies have explored the possibility of uncovering information from web search queries but few have examined the factors that affect web query data sources. We conducted a study that investigated this issue by comparing Google Trends and Baidu Index. Data from these two services are based on queries entered by users into Google and Baidu, two of the largest search engines in the world. We first compared the features and functions of the two services based on documents and extensive testing. We then carried out an empirical study that collected query volume data from the two sources. We found that data from both sources could be used to predict the quality of Chinese universities and companies. Despite the differences between the two services in terms of technology, such as differing methods of language processing, the search volume data from the two were highly correlated and combining the two data sources did not improve the predictive power of the data. However, there was a major difference between the two in terms of data availability. Baidu Index was able to provide more search volume data than Google Trends did. Our analysis showed that the disadvantage of Google Trends in this regard was due to Google's smaller user base in China. The implication of this finding goes beyond China. Google's user bases in many countries are smaller than that in China, so the search volume data related to those countries could result in the same issue as that related to China.
    Source
    Journal of the Association for Information Science and Technology. 66(2015) no.1, S.13-22
  3. Romero-Frías, E.; Vaughan, L.: Exploring the relationships between media and political parties through web hyperlink analysis : the case of Spain (2012) 0.01
    0.007566384 = product of:
      0.052964687 = sum of:
        0.052964687 = weight(_text_:case in 239) [ClassicSimilarity], result of:
          0.052964687 = score(doc=239,freq=2.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.29144385 = fieldWeight in 239, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.046875 = fieldNorm(doc=239)
      0.14285715 = coord(1/7)
    
  4. Vaughan, L.: New measurements for search engine evaluation proposed and tested (2004) 0.01
    0.005188664 = product of:
      0.036320645 = sum of:
        0.036320645 = weight(_text_:management in 2535) [ClassicSimilarity], result of:
          0.036320645 = score(doc=2535,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.2606825 = fieldWeight in 2535, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2535)
      0.14285715 = coord(1/7)
    
    Source
    Information processing and management. 40(2004) no.4, S.677-691
  5. Vaughan, L.; Thelwall, M.: Search engine coverage bias : evidence and possible causes (2004) 0.00
    0.004447426 = product of:
      0.031131983 = sum of:
        0.031131983 = weight(_text_:management in 2536) [ClassicSimilarity], result of:
          0.031131983 = score(doc=2536,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.22344214 = fieldWeight in 2536, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.046875 = fieldNorm(doc=2536)
      0.14285715 = coord(1/7)
    
    Source
    Information processing and management. 40(2004) no.4, S.693-708
  6. Vaughan, L.: Uncovering information from social media hyperlinks (2016) 0.00
    0.003672827 = product of:
      0.025709787 = sum of:
        0.025709787 = product of:
          0.051419575 = sum of:
            0.051419575 = weight(_text_:studies in 2892) [ClassicSimilarity], result of:
              0.051419575 = score(doc=2892,freq=4.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.3117402 = fieldWeight in 2892, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2892)
          0.5 = coord(1/2)
      0.14285715 = coord(1/7)
    
    Abstract
    Analyzing hyperlink patterns has been a major research topic since the early days of the web. Numerous studies reported uncovering rich information and methodological advances. However, very few studies thus far examined hyperlinks in the rapidly developing sphere of social media. This paper reports a study that helps fill this gap. The study analyzed links originating from tweets to the websites of 3 types of organizations (government, education, and business). Data were collected over an 8-month period to observe the fluctuation and reliability of the individual data set. Hyperlink data from the general web (not social media sites) were also collected and compared with social media data. The study found that the 2 types of hyperlink data correlated significantly and that analyzing the 2 together can help organizations see their relative strength or weakness in the two platforms. The study also found that both types of inlink data correlated with offline measures of organizations' performance. Twitter data from a relatively short period were fairly reliable in estimating performance measures. The timelier nature of social media data as well as the date/time stamps on tweets make this type of data potentially more valuable than that from the general web.
  7. Vaughan, L.; Ninkov, A.: ¬A new approach to web co-link analysis (2018) 0.00
    0.003672827 = product of:
      0.025709787 = sum of:
        0.025709787 = product of:
          0.051419575 = sum of:
            0.051419575 = weight(_text_:studies in 4256) [ClassicSimilarity], result of:
              0.051419575 = score(doc=4256,freq=4.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.3117402 = fieldWeight in 4256, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4256)
          0.5 = coord(1/2)
      0.14285715 = coord(1/7)
    
    Abstract
    Numerous web co-link studies have analyzed a wide variety of websites ranging from those in the academic and business arena to those dealing with politics and governments. Such studies uncover rich information about these organizations. In recent years, however, there has been a dearth of co-link analysis, mainly due to the lack of sources from which co-link data can be collected directly. Although several commercial services such as Alexa provide inlink data, none provide co-link data. We propose a new approach to web co-link analysis that can alleviate this problem so that researchers can continue to mine the valuable information contained in co-link data. The proposed approach has two components: (a) generating co-link data from inlink data using a computer program; (b) analyzing co-link data at the site level in addition to the page level that previous co-link analyses have used. The site-level analysis has the potential of expanding co-link data sources. We tested this proposed approach by analyzing a group of websites focused on vaccination using Moz inlink data. We found that the approach is feasible, as we were able to generate co-link data from inlink data and analyze the co-link data with multidimensional scaling.
  8. Thelwall, M.; Vaughan, L.; Björneborn, L.: Webometrics (2004) 0.00
    0.002597081 = product of:
      0.018179566 = sum of:
        0.018179566 = product of:
          0.03635913 = sum of:
            0.03635913 = weight(_text_:studies in 4279) [ClassicSimilarity], result of:
              0.03635913 = score(doc=4279,freq=2.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.22043361 = fieldWeight in 4279, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4279)
          0.5 = coord(1/2)
      0.14285715 = coord(1/7)
    
    Abstract
    Webometrics, the quantitative study of Web-related phenomena, emerged from the realization that methods originally designed for bibliometric analysis of scientific journal article citation patterns could be applied to the Web, with commercial search engines providing the raw data. Almind and Ingwersen (1997) defined the field and gave it its name. Other pioneers included Rodriguez Gairin (1997) and Aguillo (1998). Larson (1996) undertook exploratory link structure analysis, as did Rousseau (1997). Webometrics encompasses research from fields beyond information science such as communication studies, statistical physics, and computer science. In this review we concentrate on link analysis, but also cover other aspects of webometrics, including Web log fle analysis. One theme that runs through this chapter is the messiness of Web data and the need for data cleansing heuristics. The uncontrolled Web creates numerous problems in the interpretation of results, for instance, from the automatic creation or replication of links. The loose connection between top-level domain specifications (e.g., com, edu, and org) and their actual content is also a frustrating problem. For example, many .com sites contain noncommercial content, although com is ostensibly the main commercial top-level domain. Indeed, a skeptical researcher could claim that obstacles of this kind are so great that all Web analyses lack value. As will be seen, one response to this view, a view shared by critics of evaluative bibliometrics, is to demonstrate that Web data correlate significantly with some non-Web data in order to prove that the Web data are not wholly random. A practical response has been to develop increasingly sophisticated data cleansing techniques and multiple data analysis methods.
  9. Vaughan, L.; Yang, R.: Web data as academic and business quality estimates : a comparison of three data sources (2012) 0.00
    0.002597081 = product of:
      0.018179566 = sum of:
        0.018179566 = product of:
          0.03635913 = sum of:
            0.03635913 = weight(_text_:studies in 452) [ClassicSimilarity], result of:
              0.03635913 = score(doc=452,freq=2.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.22043361 = fieldWeight in 452, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=452)
          0.5 = coord(1/2)
      0.14285715 = coord(1/7)
    
    Abstract
    Earlier studies found that web hyperlink data contain various types of information, ranging from academic to political, that can be used to analyze a variety of social phenomena. Specifically, the numbers of inlinks to academic websites are associated with academic performance, while the counts of inlinks to company websites correlate with business variables. However, the scarcity of sources from which to collect inlink data in recent years has required us to seek new data sources. The recent demise of the inlink search function of Yahoo! made this need more pressing. Different alternative variables or data sources have been proposed. This study compared three types of web data to determine which are better as academic and business quality estimates, and what are the relationships among the three data sources. The study found that Alexa inlink and Google URL citation data can replace Yahoo! inlink data and that the former is better than the latter. Alexa is even better than Yahoo!, which has been the main data source in recent years. The unique nature of Alexa data could explain its relative advantages over other data sources.