Search (6 results, page 1 of 1)

  • × theme_ss:"Data Mining"
  • × type_ss:"a"
  • × year_i:[2010 TO 2020}
  1. Tonkin, E.L.; Tourte, G.J.L.: Working with text. tools, techniques and approaches for text mining (2016) 0.01
    0.0092614265 = product of:
      0.06482998 = sum of:
        0.06482998 = product of:
          0.12965997 = sum of:
            0.12965997 = weight(_text_:aufsatzsammlung in 4019) [ClassicSimilarity], result of:
              0.12965997 = score(doc=4019,freq=4.0), product of:
                0.25295308 = queryWeight, product of:
                  6.5610886 = idf(docFreq=169, maxDocs=44218)
                  0.038553525 = queryNorm
                0.51258504 = fieldWeight in 4019, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  6.5610886 = idf(docFreq=169, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4019)
          0.5 = coord(1/2)
      0.14285715 = coord(1/7)
    
    RSWK
    Text Mining / Aufsatzsammlung
    Subject
    Text Mining / Aufsatzsammlung
  2. Teich, E.; Degaetano-Ortlieb, S.; Fankhauser, P.; Kermes, H.; Lapshinova-Koltunski, E.: ¬The linguistic construal of disciplinarity : a data-mining approach using register features (2016) 0.00
    0.0048761885 = product of:
      0.03413332 = sum of:
        0.03413332 = weight(_text_:computer in 3015) [ClassicSimilarity], result of:
          0.03413332 = score(doc=3015,freq=2.0), product of:
            0.14089422 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.038553525 = queryNorm
            0.24226204 = fieldWeight in 3015, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.046875 = fieldNorm(doc=3015)
      0.14285715 = coord(1/7)
    
    Abstract
    We analyze the linguistic evolution of selected scientific disciplines over a 30-year time span (1970s to 2000s). Our focus is on four highly specialized disciplines at the boundaries of computer science that emerged during that time: computational linguistics, bioinformatics, digital construction, and microelectronics. Our analysis is driven by the question whether these disciplines develop a distinctive language use-both individually and collectively-over the given time period. The data set is the English Scientific Text Corpus (scitex), which includes texts from the 1970s/1980s and early 2000s. Our theoretical basis is register theory. In terms of methods, we combine corpus-based methods of feature extraction (various aggregated features [part-of-speech based], n-grams, lexico-grammatical patterns) and automatic text classification. The results of our research are directly relevant to the study of linguistic variation and languages for specific purposes (LSP) and have implications for various natural language processing (NLP) tasks, for example, authorship attribution, text mining, or training NLP tools.
  3. Hallonsten, O.; Holmberg, D.: Analyzing structural stratification in the Swedish higher education system : data contextualization with policy-history analysis (2013) 0.00
    0.0018655253 = product of:
      0.013058676 = sum of:
        0.013058676 = product of:
          0.026117353 = sum of:
            0.026117353 = weight(_text_:22 in 668) [ClassicSimilarity], result of:
              0.026117353 = score(doc=668,freq=2.0), product of:
                0.13500787 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038553525 = queryNorm
                0.19345059 = fieldWeight in 668, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=668)
          0.5 = coord(1/2)
      0.14285715 = coord(1/7)
    
    Date
    22. 3.2013 19:43:01
  4. Vaughan, L.; Chen, Y.: Data mining from web search queries : a comparison of Google trends and Baidu index (2015) 0.00
    0.0018655253 = product of:
      0.013058676 = sum of:
        0.013058676 = product of:
          0.026117353 = sum of:
            0.026117353 = weight(_text_:22 in 1605) [ClassicSimilarity], result of:
              0.026117353 = score(doc=1605,freq=2.0), product of:
                0.13500787 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038553525 = queryNorm
                0.19345059 = fieldWeight in 1605, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1605)
          0.5 = coord(1/2)
      0.14285715 = coord(1/7)
    
    Source
    Journal of the Association for Information Science and Technology. 66(2015) no.1, S.13-22
  5. Fonseca, F.; Marcinkowski, M.; Davis, C.: Cyber-human systems of thought and understanding (2019) 0.00
    0.0018655253 = product of:
      0.013058676 = sum of:
        0.013058676 = product of:
          0.026117353 = sum of:
            0.026117353 = weight(_text_:22 in 5011) [ClassicSimilarity], result of:
              0.026117353 = score(doc=5011,freq=2.0), product of:
                0.13500787 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038553525 = queryNorm
                0.19345059 = fieldWeight in 5011, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5011)
          0.5 = coord(1/2)
      0.14285715 = coord(1/7)
    
    Date
    7. 3.2019 16:32:22
  6. Jäger, L.: Von Big Data zu Big Brother (2018) 0.00
    0.0014924201 = product of:
      0.010446941 = sum of:
        0.010446941 = product of:
          0.020893881 = sum of:
            0.020893881 = weight(_text_:22 in 5234) [ClassicSimilarity], result of:
              0.020893881 = score(doc=5234,freq=2.0), product of:
                0.13500787 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.038553525 = queryNorm
                0.15476047 = fieldWeight in 5234, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5234)
          0.5 = coord(1/2)
      0.14285715 = coord(1/7)
    
    Date
    22. 1.2018 11:33:49