Search (130 results, page 1 of 7)

  • × theme_ss:"Data Mining"
  • × type_ss:"a"
  1. Fonseca, F.; Marcinkowski, M.; Davis, C.: Cyber-human systems of thought and understanding (2019) 0.05
    0.054664552 = product of:
      0.08199683 = sum of:
        0.011937855 = weight(_text_:in in 5011) [ClassicSimilarity], result of:
          0.011937855 = score(doc=5011,freq=10.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.16802745 = fieldWeight in 5011, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5011)
        0.07005897 = sum of:
          0.034676272 = weight(_text_:science in 5011) [ClassicSimilarity], result of:
            0.034676272 = score(doc=5011,freq=6.0), product of:
              0.1375819 = queryWeight, product of:
                2.6341193 = idf(docFreq=8627, maxDocs=44218)
                0.052230705 = queryNorm
              0.25204095 = fieldWeight in 5011, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                2.6341193 = idf(docFreq=8627, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5011)
          0.035382695 = weight(_text_:22 in 5011) [ClassicSimilarity], result of:
            0.035382695 = score(doc=5011,freq=2.0), product of:
              0.18290302 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052230705 = queryNorm
              0.19345059 = fieldWeight in 5011, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5011)
      0.6666667 = coord(2/3)
    
    Abstract
    The present challenge faced by scientists working with Big Data comes in the overwhelming volume and level of detail provided by current data sets. Exceeding traditional empirical approaches, Big Data opens a new perspective on scientific work in which data comes to play a role in the development of the scientific problematic to be developed. Addressing this reconfiguration of our relationship with data through readings of Wittgenstein, Macherey, and Popper, we propose a picture of science that encourages scientists to engage with the data in a direct way, using the data itself as an instrument for scientific investigation. Using GIS as a theme, we develop the concept of cyber-human systems of thought and understanding to bridge the divide between representative (theoretical) thinking and (non-theoretical) data-driven science. At the foundation of these systems, we invoke the concept of the "semantic pixel" to establish a logical and virtual space linking data and the work of scientists. It is with this discussion of the relationship between analysts in their pursuit of knowledge and the rise of Big Data that this present discussion of the philosophical foundations of Big Data addresses the central questions raised by social informatics research.
    Date
    7. 3.2019 16:32:22
    Source
    Journal of the Association for Information Science and Technology. 70(2019) no.4, S.402-411
  2. Vaughan, L.; Chen, Y.: Data mining from web search queries : a comparison of Google trends and Baidu index (2015) 0.05
    0.047002252 = product of:
      0.07050338 = sum of:
        0.015100324 = weight(_text_:in in 1605) [ClassicSimilarity], result of:
          0.015100324 = score(doc=1605,freq=16.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.21253976 = fieldWeight in 1605, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1605)
        0.05540305 = sum of:
          0.020020355 = weight(_text_:science in 1605) [ClassicSimilarity], result of:
            0.020020355 = score(doc=1605,freq=2.0), product of:
              0.1375819 = queryWeight, product of:
                2.6341193 = idf(docFreq=8627, maxDocs=44218)
                0.052230705 = queryNorm
              0.1455159 = fieldWeight in 1605, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.6341193 = idf(docFreq=8627, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1605)
          0.035382695 = weight(_text_:22 in 1605) [ClassicSimilarity], result of:
            0.035382695 = score(doc=1605,freq=2.0), product of:
              0.18290302 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052230705 = queryNorm
              0.19345059 = fieldWeight in 1605, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1605)
      0.6666667 = coord(2/3)
    
    Abstract
    Numerous studies have explored the possibility of uncovering information from web search queries but few have examined the factors that affect web query data sources. We conducted a study that investigated this issue by comparing Google Trends and Baidu Index. Data from these two services are based on queries entered by users into Google and Baidu, two of the largest search engines in the world. We first compared the features and functions of the two services based on documents and extensive testing. We then carried out an empirical study that collected query volume data from the two sources. We found that data from both sources could be used to predict the quality of Chinese universities and companies. Despite the differences between the two services in terms of technology, such as differing methods of language processing, the search volume data from the two were highly correlated and combining the two data sources did not improve the predictive power of the data. However, there was a major difference between the two in terms of data availability. Baidu Index was able to provide more search volume data than Google Trends did. Our analysis showed that the disadvantage of Google Trends in this regard was due to Google's smaller user base in China. The implication of this finding goes beyond China. Google's user bases in many countries are smaller than that in China, so the search volume data related to those countries could result in the same issue as that related to China.
    Source
    Journal of the Association for Information Science and Technology. 66(2015) no.1, S.13-22
  3. Hallonsten, O.; Holmberg, D.: Analyzing structural stratification in the Swedish higher education system : data contextualization with policy-history analysis (2013) 0.04
    0.04405373 = product of:
      0.06608059 = sum of:
        0.010677542 = weight(_text_:in in 668) [ClassicSimilarity], result of:
          0.010677542 = score(doc=668,freq=8.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.15028831 = fieldWeight in 668, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0390625 = fieldNorm(doc=668)
        0.05540305 = sum of:
          0.020020355 = weight(_text_:science in 668) [ClassicSimilarity], result of:
            0.020020355 = score(doc=668,freq=2.0), product of:
              0.1375819 = queryWeight, product of:
                2.6341193 = idf(docFreq=8627, maxDocs=44218)
                0.052230705 = queryNorm
              0.1455159 = fieldWeight in 668, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.6341193 = idf(docFreq=8627, maxDocs=44218)
                0.0390625 = fieldNorm(doc=668)
          0.035382695 = weight(_text_:22 in 668) [ClassicSimilarity], result of:
            0.035382695 = score(doc=668,freq=2.0), product of:
              0.18290302 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052230705 = queryNorm
              0.19345059 = fieldWeight in 668, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=668)
      0.6666667 = coord(2/3)
    
    Abstract
    20th century massification of higher education and research in academia is said to have produced structurally stratified higher education systems in many countries. Most manifestly, the research mission of universities appears to be divisive. Authors have claimed that the Swedish system, while formally unified, has developed into a binary state, and statistics seem to support this conclusion. This article makes use of a comprehensive statistical data source on Swedish higher education institutions to illustrate stratification, and uses literature on Swedish research policy history to contextualize the statistics. Highlighting the opportunities as well as constraints of the data, the article argues that there is great merit in combining statistics with a qualitative analysis when studying the structural characteristics of national higher education systems. Not least the article shows that it is an over-simplification to describe the Swedish system as binary; the stratification is more complex. On basis of the analysis, the article also argues that while global trends certainly influence national developments, higher education systems have country-specific features that may enrich the understanding of how systems evolve and therefore should be analyzed as part of a broader study of the increasingly globalized academic system.
    Date
    22. 3.2013 19:43:01
    Source
    Journal of the American Society for Information Science and Technology. 64(2013) no.3, S.574-586
  4. Carter, D.; Sholler, D.: Data science on the ground : hype, criticism, and everyday work (2016) 0.03
    0.030048063 = product of:
      0.045072094 = sum of:
        0.011096427 = weight(_text_:in in 3111) [ClassicSimilarity], result of:
          0.011096427 = score(doc=3111,freq=6.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.1561842 = fieldWeight in 3111, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=3111)
        0.03397567 = product of:
          0.06795134 = sum of:
            0.06795134 = weight(_text_:science in 3111) [ClassicSimilarity], result of:
              0.06795134 = score(doc=3111,freq=16.0), product of:
                0.1375819 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.052230705 = queryNorm
                0.49389738 = fieldWeight in 3111, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3111)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Modern organizations often employ data scientists to improve business processes using diverse sets of data. Researchers and practitioners have both touted the benefits and warned of the drawbacks associated with data science and big data approaches, but few studies investigate how data science is carried out "on the ground." In this paper, we first review the hype and criticisms surrounding data science and big data approaches. We then present the findings of semistructured interviews with 18 data analysts from various industries and organizational roles. Using qualitative coding techniques, we evaluated these interviews in light of the hype and criticisms surrounding data science in the popular discourse. We found that although the data analysts we interviewed were sensitive to both the allure and the potential pitfalls of data science, their motivations and evaluations of their work were more nuanced. We conclude by reflecting on the relationship between data analysts' work and the discourses around data science and big data, suggesting how future research can better account for the everyday practices of this profession.
    Source
    Journal of the Association for Information Science and Technology. 67(2016) no.10, S.2309-2319
  5. Hofstede, A.H.M. ter; Proper, H.A.; Van der Weide, T.P.: Exploiting fact verbalisation in conceptual information modelling (1997) 0.03
    0.02514248 = product of:
      0.037713718 = sum of:
        0.012945832 = weight(_text_:in in 2908) [ClassicSimilarity], result of:
          0.012945832 = score(doc=2908,freq=6.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.1822149 = fieldWeight in 2908, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2908)
        0.024767887 = product of:
          0.049535774 = sum of:
            0.049535774 = weight(_text_:22 in 2908) [ClassicSimilarity], result of:
              0.049535774 = score(doc=2908,freq=2.0), product of:
                0.18290302 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.052230705 = queryNorm
                0.2708308 = fieldWeight in 2908, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2908)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Focuses on the information modelling side of conceptual modelling. Deals with the exploitation of fact verbalisations after finishing the actual information system. Verbalisations are used as input for the design of the so-called information model. Exploits these verbalisation in 4 directions: considers their use for a conceptual query language, the verbalisation of instances, the description of the contents of a database and for the verbalisation of queries in a computer supported query environment. Provides an example session with an envisioned tool for end user query formulations that exploits the verbalisation
    Source
    Information systems. 22(1997) nos.5/6, S.349-385
  6. Bath, P.A.: Data mining in health and medical information (2003) 0.02
    0.024626605 = product of:
      0.036939908 = sum of:
        0.020923624 = weight(_text_:in in 4263) [ClassicSimilarity], result of:
          0.020923624 = score(doc=4263,freq=12.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.29450375 = fieldWeight in 4263, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=4263)
        0.016016284 = product of:
          0.032032568 = sum of:
            0.032032568 = weight(_text_:science in 4263) [ClassicSimilarity], result of:
              0.032032568 = score(doc=4263,freq=2.0), product of:
                0.1375819 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.052230705 = queryNorm
                0.23282544 = fieldWeight in 4263, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4263)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Data mining (DM) is part of a process by which information can be extracted from data or databases and used to inform decision making in a variety of contexts (Benoit, 2002; Michalski, Bratka & Kubat, 1997). DM includes a range of tools and methods for extractiog information; their use in the commercial sector for knowledge extraction and discovery has been one of the main driving forces in their development (Adriaans & Zantinge, 1996; Benoit, 2002). DM has been developed and applied in numerous areas. This review describes its use in analyzing health and medical information.
    Source
    Annual review of information science and technology. 38(2004), S.331-370
  7. Matson, L.D.; Bonski, D.J.: Do digital libraries need librarians? (1997) 0.02
    0.024565458 = product of:
      0.036848187 = sum of:
        0.008542033 = weight(_text_:in in 1737) [ClassicSimilarity], result of:
          0.008542033 = score(doc=1737,freq=2.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.120230645 = fieldWeight in 1737, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=1737)
        0.028306156 = product of:
          0.056612313 = sum of:
            0.056612313 = weight(_text_:22 in 1737) [ClassicSimilarity], result of:
              0.056612313 = score(doc=1737,freq=2.0), product of:
                0.18290302 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.052230705 = queryNorm
                0.30952093 = fieldWeight in 1737, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1737)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Defines digital libraries and discusses the effects of new technology on librarians. Examines the different viewpoints of librarians and information technologists on digital libraries. Describes the development of a digital library at the National Drug Intelligence Center, USA, which was carried out in collaboration with information technology experts. The system is based on Web enabled search technology to find information, data visualization and data mining to visualize it and use of SGML as an information standard to store it
    Date
    22.11.1998 18:57:22
  8. Amir, A.; Feldman, R.; Kashi, R.: ¬A new and versatile method for association generation (1997) 0.02
    0.024565458 = product of:
      0.036848187 = sum of:
        0.008542033 = weight(_text_:in in 1270) [ClassicSimilarity], result of:
          0.008542033 = score(doc=1270,freq=2.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.120230645 = fieldWeight in 1270, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=1270)
        0.028306156 = product of:
          0.056612313 = sum of:
            0.056612313 = weight(_text_:22 in 1270) [ClassicSimilarity], result of:
              0.056612313 = score(doc=1270,freq=2.0), product of:
                0.18290302 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.052230705 = queryNorm
                0.30952093 = fieldWeight in 1270, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1270)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Current algorithms for finding associations among the attributes describing data in a database have a number of shortcomings. Presents a novel method for association generation, that answers all desiderata. The method is different from all existing algorithms and especially suitable to textual databases with binary attributes. Uses subword trees for quick indexing into the required database statistics. Tests the algorithm on the Reuters-22173 database with satisfactory results
    Source
    Information systems. 22(1997) nos.5/6, S.333-347
  9. Frické, M.: Big data and its epistemology (2015) 0.02
    0.023946881 = product of:
      0.035920322 = sum of:
        0.009060195 = weight(_text_:in in 1811) [ClassicSimilarity], result of:
          0.009060195 = score(doc=1811,freq=4.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.12752387 = fieldWeight in 1811, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=1811)
        0.026860125 = product of:
          0.05372025 = sum of:
            0.05372025 = weight(_text_:science in 1811) [ClassicSimilarity], result of:
              0.05372025 = score(doc=1811,freq=10.0), product of:
                0.1375819 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.052230705 = queryNorm
                0.39046016 = fieldWeight in 1811, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1811)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The article considers whether Big Data, in the form of data-driven science, will enable the discovery, or appraisal, of universal scientific theories, instrumentalist tools, or inductive inferences. It points out, initially, that such aspirations are similar to the now-discredited inductivist approach to science. On the positive side, Big Data may permit larger sample sizes, cheaper and more extensive testing of theories, and the continuous assessment of theories. On the negative side, data-driven science encourages passive data collection, as opposed to experimentation and testing, and hornswoggling ("unsound statistical fiddling"). The roles of theory and data in inductive algorithms, statistical modeling, and scientific discoveries are analyzed, and it is argued that theory is needed at every turn. Data-driven science is a chimera.
    Source
    Journal of the Association for Information Science and Technology. 66(2015) no.4, S.651-661
  10. Leydesdorff, L.; Persson, O.: Mapping the geography of science : distribution patterns and networks of relations among cities and institutes (2010) 0.02
    0.022056414 = product of:
      0.03308462 = sum of:
        0.009060195 = weight(_text_:in in 3704) [ClassicSimilarity], result of:
          0.009060195 = score(doc=3704,freq=4.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.12752387 = fieldWeight in 3704, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=3704)
        0.024024425 = product of:
          0.04804885 = sum of:
            0.04804885 = weight(_text_:science in 3704) [ClassicSimilarity], result of:
              0.04804885 = score(doc=3704,freq=8.0), product of:
                0.1375819 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.052230705 = queryNorm
                0.34923816 = fieldWeight in 3704, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3704)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Using Google Earth, Google Maps, and/or network visualization programs such as Pajek, one can overlay the network of relations among addresses in scientific publications onto the geographic map. The authors discuss the pros and cons of various options, and provide software (freeware) for bridging existing gaps between the Science Citation Indices (Thomson Reuters) and Scopus (Elsevier), on the one hand, and these various visualization tools on the other. At the level of city names, the global map can be drawn reliably on the basis of the available address information. At the level of the names of organizations and institutes, there are problems of unification both in the ISI databases and with Scopus. Pajek enables a combination of visualization and statistical analysis, whereas the Google Maps and its derivatives provide superior tools on the Internet.
    Object
    Science Citation Index
    Source
    Journal of the American Society for Information Science and Technology. 61(2010) no.8, S.1622-1634
  11. Thelwall, M.; Wilkinson, D.; Uppal, S.: Data mining emotion in social network communication : gender differences in MySpace (2009) 0.02
    0.02087551 = product of:
      0.031313263 = sum of:
        0.014325427 = weight(_text_:in in 3322) [ClassicSimilarity], result of:
          0.014325427 = score(doc=3322,freq=10.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.20163295 = fieldWeight in 3322, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=3322)
        0.016987834 = product of:
          0.03397567 = sum of:
            0.03397567 = weight(_text_:science in 3322) [ClassicSimilarity], result of:
              0.03397567 = score(doc=3322,freq=4.0), product of:
                0.1375819 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.052230705 = queryNorm
                0.24694869 = fieldWeight in 3322, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3322)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Despite the rapid growth in social network sites and in data mining for emotion (sentiment analysis), little research has tied the two together, and none has had social science goals. This article examines the extent to which emotion is present in MySpace comments, using a combination of data mining and content analysis, and exploring age and gender. A random sample of 819 public comments to or from U.S. users was manually classified for strength of positive and negative emotion. Two thirds of the comments expressed positive emotion, but a minority (20%) contained negative emotion, confirming that MySpace is an extraordinarily emotion-rich environment. Females are likely to give and receive more positive comments than are males, but there is no difference for negative comments. It is thus possible that females are more successful social network site users partly because of their greater ability to textually harness positive affect.
    Source
    Journal of the American Society for Information Science and Technology. 61(2010) no.1, S.190-199
  12. Qiu, X.Y.; Srinivasan, P.; Hu, Y.: Supervised learning models to predict firm performance with annual reports : an empirical study (2014) 0.02
    0.020821191 = product of:
      0.031231787 = sum of:
        0.019219575 = weight(_text_:in in 1205) [ClassicSimilarity], result of:
          0.019219575 = score(doc=1205,freq=18.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.27051896 = fieldWeight in 1205, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=1205)
        0.012012213 = product of:
          0.024024425 = sum of:
            0.024024425 = weight(_text_:science in 1205) [ClassicSimilarity], result of:
              0.024024425 = score(doc=1205,freq=2.0), product of:
                0.1375819 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.052230705 = queryNorm
                0.17461908 = fieldWeight in 1205, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1205)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Text mining and machine learning methodologies have been applied toward knowledge discovery in several domains, such as biomedicine and business. Interestingly, in the business domain, the text mining and machine learning community has minimally explored company annual reports with their mandatory disclosures. In this study, we explore the question "How can annual reports be used to predict change in company performance from one year to the next?" from a text mining perspective. Our article contributes a systematic study of the potential of company mandatory disclosures using a computational viewpoint in the following aspects: (a) We characterize our research problem along distinct dimensions to gain a reasonably comprehensive understanding of the capacity of supervised learning methods in predicting change in company performance using annual reports, and (b) our findings from unbiased systematic experiments provide further evidence about the economic incentives faced by analysts in their stock recommendations and speculations on analysts having access to more information in producing earnings forecast.
    Source
    Journal of the Association for Information Science and Technology. 65(2014) no.2, S.400-413
  13. Liu, X.; Yu, S.; Janssens, F.; Glänzel, W.; Moreau, Y.; Moor, B.de: Weighted hybrid clustering by combining text mining and bibliometrics on a large-scale journal database (2010) 0.02
    0.019867256 = product of:
      0.029800884 = sum of:
        0.012813049 = weight(_text_:in in 3464) [ClassicSimilarity], result of:
          0.012813049 = score(doc=3464,freq=8.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.18034597 = fieldWeight in 3464, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=3464)
        0.016987834 = product of:
          0.03397567 = sum of:
            0.03397567 = weight(_text_:science in 3464) [ClassicSimilarity], result of:
              0.03397567 = score(doc=3464,freq=4.0), product of:
                0.1375819 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.052230705 = queryNorm
                0.24694869 = fieldWeight in 3464, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3464)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    We propose a new hybrid clustering framework to incorporate text mining with bibliometrics in journal set analysis. The framework integrates two different approaches: clustering ensemble and kernel-fusion clustering. To improve the flexibility and the efficiency of processing large-scale data, we propose an information-based weighting scheme to leverage the effect of multiple data sources in hybrid clustering. Three different algorithms are extended by the proposed weighting scheme and they are employed on a large journal set retrieved from the Web of Science (WoS) database. The clustering performance of the proposed algorithms is systematically evaluated using multiple evaluation methods, and they were cross-compared with alternative methods. Experimental results demonstrate that the proposed weighted hybrid clustering strategy is superior to other methods in clustering performance and efficiency. The proposed approach also provides a more refined structural mapping of journal sets, which is useful for monitoring and detecting new trends in different scientific fields.
    Source
    Journal of the American Society for Information Science and Technology. 61(2010) no.6, S.1105-1119
  14. Borgman, C.L.; Wofford, M.F.; Golshan, M.S.; Darch, P.T.: Collaborative qualitative research at scale : reflections on 20 years of acquiring global data and making data global (2021) 0.02
    0.019517329 = product of:
      0.029275991 = sum of:
        0.011937855 = weight(_text_:in in 239) [ClassicSimilarity], result of:
          0.011937855 = score(doc=239,freq=10.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.16802745 = fieldWeight in 239, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0390625 = fieldNorm(doc=239)
        0.017338136 = product of:
          0.034676272 = sum of:
            0.034676272 = weight(_text_:science in 239) [ClassicSimilarity], result of:
              0.034676272 = score(doc=239,freq=6.0), product of:
                0.1375819 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.052230705 = queryNorm
                0.25204095 = fieldWeight in 239, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=239)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    A 5-year project to study scientific data uses in geography, starting in 1999, evolved into 20 years of research on data practices in sensor networks, environmental sciences, biology, seismology, undersea science, biomedicine, astronomy, and other fields. By emulating the "team science" approaches of the scientists studied, the UCLA Center for Knowledge Infrastructures accumulated a comprehensive collection of qualitative data about how scientists generate, manage, use, and reuse data across domains. Building upon Paul N. Edwards's model of "making global data"-collecting signals via consistent methods, technologies, and policies-to "make data global"-comparing and integrating those data, the research team has managed and exploited these data as a collaborative resource. This article reflects on the social, technical, organizational, economic, and policy challenges the team has encountered in creating new knowledge from data old and new. We reflect on continuity over generations of students and staff, transitions between grants, transfer of legacy data between software tools, research methods, and the role of professional data managers in the social sciences.
    Source
    Journal of the Association for Information Science and Technology. 72(2021) no.6, S.667-682
  15. Bell, D.A.; Guan, J.W.: Computational methods for rough classification and discovery (1998) 0.02
    0.019308537 = product of:
      0.028962806 = sum of:
        0.014948557 = weight(_text_:in in 2909) [ClassicSimilarity], result of:
          0.014948557 = score(doc=2909,freq=8.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.21040362 = fieldWeight in 2909, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2909)
        0.014014249 = product of:
          0.028028497 = sum of:
            0.028028497 = weight(_text_:science in 2909) [ClassicSimilarity], result of:
              0.028028497 = score(doc=2909,freq=2.0), product of:
                0.1375819 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.052230705 = queryNorm
                0.20372227 = fieldWeight in 2909, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2909)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Rough set theory is a mathematical tool to deal with vagueness and uncertainty. To apply the theory, it needs to be associated with efficient and effective computational methods. A relation can be used to represent a decison table for use in decision making. By using this kind of table, rough set theory can be applied successfully to rough classification and knowledge discovery. Presents computational methods for using rough sets to identify classes in datasets, finding dependencies in relations, and discovering rules which are hidden in databases. Illustrates the methods with a running example from a database of car test results
    Source
    Journal of the American Society for Information Science. 49(1998) no.5, S.403-414
  16. Lam, W.; Yang, C.C.; Menczer, F.: Introduction to the special topic section on mining Web resources for enhancing information retrieval (2007) 0.02
    0.019308537 = product of:
      0.028962806 = sum of:
        0.014948557 = weight(_text_:in in 600) [ClassicSimilarity], result of:
          0.014948557 = score(doc=600,freq=8.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.21040362 = fieldWeight in 600, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0546875 = fieldNorm(doc=600)
        0.014014249 = product of:
          0.028028497 = sum of:
            0.028028497 = weight(_text_:science in 600) [ClassicSimilarity], result of:
              0.028028497 = score(doc=600,freq=2.0), product of:
                0.1375819 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.052230705 = queryNorm
                0.20372227 = fieldWeight in 600, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=600)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The amount of information on the Web has been expanding at an enormous pace. There are a variety of Web documents in different genres, such as news, reports, reviews. Traditionally, the information displayed on Web sites has been static. Recently, there are many Web sites offering content that is dynamically generated and frequently updated. It is also common for Web sites to contain information in different languages since many countries adopt more than one language. Moreover, content may exist in multimedia formats including text, images, video, and audio.
    Footnote
    Einführung in einen Themenschwerpunkt "Mining Web resources for enhancing information retrieval"
    Source
    Journal of the American Society for Information Science and Technology. 58(2007) no.12, S.1791-1792
  17. Schwartz, F.; Fang, Y.C.: Citation data analysis on hydrogeology (2007) 0.02
    0.018939527 = product of:
      0.02840929 = sum of:
        0.017084066 = weight(_text_:in in 433) [ClassicSimilarity], result of:
          0.017084066 = score(doc=433,freq=32.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.24046129 = fieldWeight in 433, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.03125 = fieldNorm(doc=433)
        0.011325223 = product of:
          0.022650447 = sum of:
            0.022650447 = weight(_text_:science in 433) [ClassicSimilarity], result of:
              0.022650447 = score(doc=433,freq=4.0), product of:
                0.1375819 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.052230705 = queryNorm
                0.16463245 = fieldWeight in 433, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.03125 = fieldNorm(doc=433)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This article explores the status of research in hydrogeology using data mining techniques. First we try to explain what citation analysis is and review some of the previous work on citation analysis. The main idea in this article is to address some common issues about citation numbers and the use of these data. To validate the use of citation numbers, we compare the citation patterns for Water Resources Research papers in the 1980s with those in the 1990s. The citation growths for highly cited authors from the 1980s are used to examine whether it is possible to predict the citation patterns for highly-cited authors in the 1990s. If the citation data prove to be steady and stable, these numbers then can be used to explore the evolution of science in hydrogeology. The famous quotation, "If you are not the lead dog, the scenery never changes," attributed to Lee Iacocca, points to the importance of an entrepreneurial spirit in all forms of endeavor. In the case of hydrogeological research, impact analysis makes it clear how important it is to be a pioneer. Statistical correlation coefficients are used to retrieve papers among a collection of 2,847 papers before and after 1991 sharing the same topics with 273 papers in 1991 in Water Resources Research. The numbers of papers before and after 1991 are then plotted against various levels of citations for papers in 1991 to compare the distributions of paper population before and after that year. The similarity metrics based on word counts can ensure that the "before" papers are like ancestors and "after" papers are descendants in the same type of research. This exercise gives us an idea of how many papers are populated before and after 1991 (1991 is chosen based on balanced numbers of papers before and after that year). In addition, the impact of papers is measured in terms of citation presented as "percentile," a relative measure based on rankings in one year, in order to minimize the effect of time.
    Source
    Journal of the American Society for Information Science and Technology. 58(2007) no.4, S.518-525
  18. Derek Doran, D.; Gokhale, S.S.: ¬A classification framework for web robots (2012) 0.02
    0.01873103 = product of:
      0.028096544 = sum of:
        0.01208026 = weight(_text_:in in 505) [ClassicSimilarity], result of:
          0.01208026 = score(doc=505,freq=4.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.17003182 = fieldWeight in 505, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0625 = fieldNorm(doc=505)
        0.016016284 = product of:
          0.032032568 = sum of:
            0.032032568 = weight(_text_:science in 505) [ClassicSimilarity], result of:
              0.032032568 = score(doc=505,freq=2.0), product of:
                0.1375819 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.052230705 = queryNorm
                0.23282544 = fieldWeight in 505, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0625 = fieldNorm(doc=505)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The behavior of modern web robots varies widely when they crawl for different purposes. In this article, we present a framework to classify these web robots from two orthogonal perspectives, namely, their functionality and the types of resources they consume. Applying the classification framework to a year-long access log from the UConn SoE web server, we present trends that point to significant differences in their crawling behavior.
    Source
    Journal of the American Society for Information Science and Technology. 63(2012) no.12, S.2549-2554,
  19. Song, J.; Huang, Y.; Qi, X.; Li, Y.; Li, F.; Fu, K.; Huang, T.: Discovering hierarchical topic evolution in time-stamped documents (2016) 0.02
    0.018469956 = product of:
      0.027704932 = sum of:
        0.015692718 = weight(_text_:in in 2853) [ClassicSimilarity], result of:
          0.015692718 = score(doc=2853,freq=12.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.22087781 = fieldWeight in 2853, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=2853)
        0.012012213 = product of:
          0.024024425 = sum of:
            0.024024425 = weight(_text_:science in 2853) [ClassicSimilarity], result of:
              0.024024425 = score(doc=2853,freq=2.0), product of:
                0.1375819 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.052230705 = queryNorm
                0.17461908 = fieldWeight in 2853, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2853)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The objective of this paper is to propose a hierarchical topic evolution model (HTEM) that can organize time-varying topics in a hierarchy and discover their evolutions with multiple timescales. In the proposed HTEM, topics near the root of the hierarchy are more abstract and also evolve in the longer timescales than those near the leaves. To achieve this goal, the distance-dependent Chinese restaurant process (ddCRP) is extended to a new nested process that is able to simultaneously model the dependencies among data and the relationship between clusters. The HTEM is proposed based on the new process for time-stamped documents, in which the timestamp is utilized to measure the dependencies among documents. Moreover, an efficient Gibbs sampler is developed for the proposed HTEM. Our experimental results on two popular real-world data sets verify that the proposed HTEM can capture coherent topics and discover their hierarchical evolutions. It also outperforms the baseline model in terms of likelihood on held-out data.
    Source
    Journal of the Association for Information Science and Technology. 67(2016) no.4, S.915-927
  20. Ekbia, H.; Mattioli, M.; Kouper, I.; Arave, G.; Ghazinejad, A.; Bowman, T.; Suri, V.R.; Tsou, A.; Weingart, S.; Sugimoto, C.R.: Big data, bigger dilemmas : a critical review (2015) 0.02
    0.018155862 = product of:
      0.027233792 = sum of:
        0.013077264 = weight(_text_:in in 2155) [ClassicSimilarity], result of:
          0.013077264 = score(doc=2155,freq=12.0), product of:
            0.07104705 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.052230705 = queryNorm
            0.18406484 = fieldWeight in 2155, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2155)
        0.014156529 = product of:
          0.028313057 = sum of:
            0.028313057 = weight(_text_:science in 2155) [ClassicSimilarity], result of:
              0.028313057 = score(doc=2155,freq=4.0), product of:
                0.1375819 = queryWeight, product of:
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.052230705 = queryNorm
                0.20579056 = fieldWeight in 2155, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.6341193 = idf(docFreq=8627, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2155)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The recent interest in Big Data has generated a broad range of new academic, corporate, and policy practices along with an evolving debate among its proponents, detractors, and skeptics. While the practices draw on a common set of tools, techniques, and technologies, most contributions to the debate come either from a particular disciplinary perspective or with a focus on a domain-specific issue. A close examination of these contributions reveals a set of common problematics that arise in various guises and in different places. It also demonstrates the need for a critical synthesis of the conceptual and practical dilemmas surrounding Big Data. The purpose of this article is to provide such a synthesis by drawing on relevant writings in the sciences, humanities, policy, and trade literature. In bringing these diverse literatures together, we aim to shed light on the common underlying issues that concern and affect all of these areas. By contextualizing the phenomenon of Big Data within larger socioeconomic developments, we also seek to provide a broader understanding of its drivers, barriers, and challenges. This approach allows us to identify attributes of Big Data that require more attention-autonomy, opacity, generativity, disparity, and futurity-leading to questions and ideas for moving beyond dilemmas.
    Series
    Advances in information science
    Source
    Journal of the Association for Information Science and Technology. 66(2015) no.8, S.1523-1545

Years

Languages

  • e 109
  • d 21

Classifications