Search (25 results, page 1 of 2)

  • × theme_ss:"Data Mining"
  1. Mining text data (2012) 0.02
    0.015501413 = product of:
      0.046504237 = sum of:
        0.046504237 = product of:
          0.09300847 = sum of:
            0.09300847 = weight(_text_:networks in 362) [ClassicSimilarity], result of:
              0.09300847 = score(doc=362,freq=8.0), product of:
                0.22247115 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.047034867 = queryNorm
                0.4180698 = fieldWeight in 362, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.03125 = fieldNorm(doc=362)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Text mining applications have experienced tremendous advances because of web 2.0 and social networking applications. Recent advances in hardware and software technology have lead to a number of unique scenarios where text mining algorithms are learned. Mining Text Data introduces an important niche in the text analytics field, and is an edited volume contributed by leading international researchers and practitioners focused on social networks & data mining. This book contains a wide swath in topics across social networks & data mining. Each chapter contains a comprehensive survey including the key research content on the topic, and the future directions of research in the field. There is a special focus on Text Embedded with Heterogeneous and Multimedia Data which makes the mining process much more challenging. A number of methods have been designed such as transfer learning and cross-lingual mining for such cases. Mining Text Data simplifies the content, so that advanced-level students, practitioners and researchers in computer science can benefit from this book. Academic and corporate libraries, as well as ACM, IEEE, and Management Science focused on information security, electronic commerce, databases, data mining, machine learning, and statistics are the primary buyers for this reference book.
    LCSH
    Computer Communication Networks
    Subject
    Computer Communication Networks
  2. Chowdhury, G.G.: Template mining for information extraction from digital documents (1999) 0.01
    0.014869343 = product of:
      0.044608027 = sum of:
        0.044608027 = product of:
          0.08921605 = sum of:
            0.08921605 = weight(_text_:22 in 4577) [ClassicSimilarity], result of:
              0.08921605 = score(doc=4577,freq=2.0), product of:
                0.1647081 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047034867 = queryNorm
                0.5416616 = fieldWeight in 4577, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4577)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    2. 4.2000 18:01:22
  3. Methodologies for knowledge discovery and data mining : Third Pacific-Asia Conference, PAKDD'99, Beijing, China, April 26-28, 1999, Proceedings (1999) 0.01
    0.013563735 = product of:
      0.040691204 = sum of:
        0.040691204 = product of:
          0.08138241 = sum of:
            0.08138241 = weight(_text_:networks in 3821) [ClassicSimilarity], result of:
              0.08138241 = score(doc=3821,freq=2.0), product of:
                0.22247115 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.047034867 = queryNorm
                0.36581108 = fieldWeight in 3821, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3821)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    The 29 revised full papers presented together with 37 short papers were carefully selected from a total of 158 submissions. The book is divided into sections on emerging KDD technology; association rules; feature selection and generation; mining in semi-unstructured data; interestingness, surprisingness, and exceptions; rough sets, fuzzy logic, and neural networks; induction, classification, and clustering; visualization, causal models and graph-based methods; agent-based and distributed data mining; and advanced topics and new methodologies
  4. KDD : techniques and applications (1998) 0.01
    0.012745151 = product of:
      0.038235452 = sum of:
        0.038235452 = product of:
          0.076470904 = sum of:
            0.076470904 = weight(_text_:22 in 6783) [ClassicSimilarity], result of:
              0.076470904 = score(doc=6783,freq=2.0), product of:
                0.1647081 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047034867 = queryNorm
                0.46428138 = fieldWeight in 6783, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6783)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Footnote
    A special issue of selected papers from the Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD'97), held Singapore, 22-23 Feb 1997
  5. Whittle, M.; Eaglestone, B.; Ford, N.; Gillet, V.J.; Madden, A.: Data mining of search engine logs (2007) 0.01
    0.011626059 = product of:
      0.034878176 = sum of:
        0.034878176 = product of:
          0.06975635 = sum of:
            0.06975635 = weight(_text_:networks in 1330) [ClassicSimilarity], result of:
              0.06975635 = score(doc=1330,freq=2.0), product of:
                0.22247115 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.047034867 = queryNorm
                0.31355235 = fieldWeight in 1330, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1330)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    This article reports on the development of a novel method for the analysis of Web logs. The method uses techniques that look for similarities between queries and identify sequences of query transformation. It allows sequences of query transformations to be represented as graphical networks, thereby giving a richer view of search behavior than is possible with the usual sequential descriptions. We also perform a basic analysis to study the correlations between observed transformation codes, with results that appear to show evidence of behavior habits. The method was developed using transaction logs from the Excite search engine to provide a tool for an ongoing research project that is endeavoring to develop a greater understanding of Web-based searching by the general public.
  6. Leydesdorff, L.; Persson, O.: Mapping the geography of science : distribution patterns and networks of relations among cities and institutes (2010) 0.01
    0.011626059 = product of:
      0.034878176 = sum of:
        0.034878176 = product of:
          0.06975635 = sum of:
            0.06975635 = weight(_text_:networks in 3704) [ClassicSimilarity], result of:
              0.06975635 = score(doc=3704,freq=2.0), product of:
                0.22247115 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.047034867 = queryNorm
                0.31355235 = fieldWeight in 3704, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3704)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
  7. Berendt, B.; Krause, B.; Kolbe-Nusser, S.: Intelligent scientific authoring tools : interactive data mining for constructive uses of citation networks (2010) 0.01
    0.011626059 = product of:
      0.034878176 = sum of:
        0.034878176 = product of:
          0.06975635 = sum of:
            0.06975635 = weight(_text_:networks in 4226) [ClassicSimilarity], result of:
              0.06975635 = score(doc=4226,freq=2.0), product of:
                0.22247115 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.047034867 = queryNorm
                0.31355235 = fieldWeight in 4226, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4226)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
  8. Nicholson, S.: Bibliomining for automated collection development in a digital library setting : using data mining to discover Web-based scholarly research works (2003) 0.01
    0.009688382 = product of:
      0.029065145 = sum of:
        0.029065145 = product of:
          0.05813029 = sum of:
            0.05813029 = weight(_text_:networks in 1867) [ClassicSimilarity], result of:
              0.05813029 = score(doc=1867,freq=2.0), product of:
                0.22247115 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.047034867 = queryNorm
                0.26129362 = fieldWeight in 1867, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1867)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    This research creates an intelligent agent for automated collection development in a digital library setting. It uses a predictive model based an facets of each Web page to select scholarly works. The criteria came from the academic library selection literature, and a Delphi study was used to refine the list to 41 criteria. A Perl program was designed to analyze a Web page for each criterion and applied to a large collection of scholarly and nonscholarly Web pages. Bibliomining, or data mining for libraries, was then used to create different classification models. Four techniques were used: logistic regression, nonparametric discriminant analysis, classification trees, and neural networks. Accuracy and return were used to judge the effectiveness of each model an test datasets. In addition, a set of problematic pages that were difficult to classify because of their similarity to scholarly research was gathered and classified using the models. The resulting models could be used in the selection process to automatically create a digital library of Webbased scholarly research works. In addition, the technique can be extended to create a digital library of any type of structured electronic information.
  9. Haravu, L.J.; Neelameghan, A.: Text mining and data mining in knowledge organization and discovery : the making of knowledge-based products (2003) 0.01
    0.009688382 = product of:
      0.029065145 = sum of:
        0.029065145 = product of:
          0.05813029 = sum of:
            0.05813029 = weight(_text_:networks in 5653) [ClassicSimilarity], result of:
              0.05813029 = score(doc=5653,freq=2.0), product of:
                0.22247115 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.047034867 = queryNorm
                0.26129362 = fieldWeight in 5653, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5653)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Discusses the importance of knowledge organization in the context of the information overload caused by the vast quantities of data and information accessible on internal and external networks of an organization. Defines the characteristics of a knowledge-based product. Elaborates on the techniques and applications of text mining in developing knowledge products. Presents two approaches, as case studies, to the making of knowledge products: (1) steps and processes in the planning, designing and development of a composite multilingual multimedia CD product, with the potential international, inter-cultural end users in view, and (2) application of natural language processing software in text mining. Using a text mining software, it is possible to link concept terms from a processed text to a related thesaurus, glossary, schedules of a classification scheme, and facet structured subject representations. Concludes that the products of text mining and data mining could be made more useful if the features of a faceted scheme for subject classification are incorporated into text mining techniques and products.
  10. Borgman, C.L.; Wofford, M.F.; Golshan, M.S.; Darch, P.T.: Collaborative qualitative research at scale : reflections on 20 years of acquiring global data and making data global (2021) 0.01
    0.009688382 = product of:
      0.029065145 = sum of:
        0.029065145 = product of:
          0.05813029 = sum of:
            0.05813029 = weight(_text_:networks in 239) [ClassicSimilarity], result of:
              0.05813029 = score(doc=239,freq=2.0), product of:
                0.22247115 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.047034867 = queryNorm
                0.26129362 = fieldWeight in 239, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=239)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    A 5-year project to study scientific data uses in geography, starting in 1999, evolved into 20 years of research on data practices in sensor networks, environmental sciences, biology, seismology, undersea science, biomedicine, astronomy, and other fields. By emulating the "team science" approaches of the scientists studied, the UCLA Center for Knowledge Infrastructures accumulated a comprehensive collection of qualitative data about how scientists generate, manage, use, and reuse data across domains. Building upon Paul N. Edwards's model of "making global data"-collecting signals via consistent methods, technologies, and policies-to "make data global"-comparing and integrating those data, the research team has managed and exploited these data as a collaborative resource. This article reflects on the social, technical, organizational, economic, and policy challenges the team has encountered in creating new knowledge from data old and new. We reflect on continuity over generations of students and staff, transitions between grants, transfer of legacy data between software tools, research methods, and the role of professional data managers in the social sciences.
  11. Matson, L.D.; Bonski, D.J.: Do digital libraries need librarians? (1997) 0.01
    0.008496767 = product of:
      0.0254903 = sum of:
        0.0254903 = product of:
          0.0509806 = sum of:
            0.0509806 = weight(_text_:22 in 1737) [ClassicSimilarity], result of:
              0.0509806 = score(doc=1737,freq=2.0), product of:
                0.1647081 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047034867 = queryNorm
                0.30952093 = fieldWeight in 1737, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1737)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22.11.1998 18:57:22
  12. Lusti, M.: Data Warehousing and Data Mining : Eine Einführung in entscheidungsunterstützende Systeme (1999) 0.01
    0.008496767 = product of:
      0.0254903 = sum of:
        0.0254903 = product of:
          0.0509806 = sum of:
            0.0509806 = weight(_text_:22 in 4261) [ClassicSimilarity], result of:
              0.0509806 = score(doc=4261,freq=2.0), product of:
                0.1647081 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047034867 = queryNorm
                0.30952093 = fieldWeight in 4261, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4261)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    17. 7.2002 19:22:06
  13. Amir, A.; Feldman, R.; Kashi, R.: ¬A new and versatile method for association generation (1997) 0.01
    0.008496767 = product of:
      0.0254903 = sum of:
        0.0254903 = product of:
          0.0509806 = sum of:
            0.0509806 = weight(_text_:22 in 1270) [ClassicSimilarity], result of:
              0.0509806 = score(doc=1270,freq=2.0), product of:
                0.1647081 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047034867 = queryNorm
                0.30952093 = fieldWeight in 1270, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1270)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Source
    Information systems. 22(1997) nos.5/6, S.333-347
  14. Kantardzic, M.: Data mining : concepts, models, methods, and algorithms (2003) 0.01
    0.0077507063 = product of:
      0.023252118 = sum of:
        0.023252118 = product of:
          0.046504237 = sum of:
            0.046504237 = weight(_text_:networks in 2291) [ClassicSimilarity], result of:
              0.046504237 = score(doc=2291,freq=2.0), product of:
                0.22247115 = queryWeight, product of:
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.047034867 = queryNorm
                0.2090349 = fieldWeight in 2291, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.72992 = idf(docFreq=1060, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2291)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    This book offers a comprehensive introduction to the exploding field of data mining. We are surrounded by data, numerical and otherwise, which must be analyzed and processed to convert it into information that informs, instructs, answers, or otherwise aids understanding and decision-making. Due to the ever-increasing complexity and size of today's data sets, a new term, data mining, was created to describe the indirect, automatic data analysis techniques that utilize more complex and sophisticated tools than those which analysts used in the past to do mere data analysis. "Data Mining: Concepts, Models, Methods, and Algorithms" discusses data mining principles and then describes representative state-of-the-art methods and algorithms originating from different disciplines such as statistics, machine learning, neural networks, fuzzy logic, and evolutionary computation. Detailed algorithms are provided with necessary explanations and illustrative examples. This text offers guidance: how and when to use a particular software tool (with their companion data sets) from among the hundreds offered when faced with a data set to mine. This allows analysts to create and perform their own data mining experiments using their knowledge of the methodologies and techniques provided. This book emphasizes the selection of appropriate methodologies and data analysis software, as well as parameter tuning. These critically important, qualitative decisions can only be made with the deeper understanding of parameter meaning and its role in the technique that is offered here. Data mining is an exploding field and this book offers much-needed guidance to selecting among the numerous analysis programs that are available.
  15. Hofstede, A.H.M. ter; Proper, H.A.; Van der Weide, T.P.: Exploiting fact verbalisation in conceptual information modelling (1997) 0.01
    0.0074346713 = product of:
      0.022304013 = sum of:
        0.022304013 = product of:
          0.044608027 = sum of:
            0.044608027 = weight(_text_:22 in 2908) [ClassicSimilarity], result of:
              0.044608027 = score(doc=2908,freq=2.0), product of:
                0.1647081 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047034867 = queryNorm
                0.2708308 = fieldWeight in 2908, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2908)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Source
    Information systems. 22(1997) nos.5/6, S.349-385
  16. Lackes, R.; Tillmanns, C.: Data Mining für die Unternehmenspraxis : Entscheidungshilfen und Fallstudien mit führenden Softwarelösungen (2006) 0.01
    0.0063725756 = product of:
      0.019117726 = sum of:
        0.019117726 = product of:
          0.038235452 = sum of:
            0.038235452 = weight(_text_:22 in 1383) [ClassicSimilarity], result of:
              0.038235452 = score(doc=1383,freq=2.0), product of:
                0.1647081 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047034867 = queryNorm
                0.23214069 = fieldWeight in 1383, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1383)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 3.2008 14:46:06
  17. Hallonsten, O.; Holmberg, D.: Analyzing structural stratification in the Swedish higher education system : data contextualization with policy-history analysis (2013) 0.01
    0.0053104796 = product of:
      0.015931439 = sum of:
        0.015931439 = product of:
          0.031862877 = sum of:
            0.031862877 = weight(_text_:22 in 668) [ClassicSimilarity], result of:
              0.031862877 = score(doc=668,freq=2.0), product of:
                0.1647081 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047034867 = queryNorm
                0.19345059 = fieldWeight in 668, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=668)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 3.2013 19:43:01
  18. Vaughan, L.; Chen, Y.: Data mining from web search queries : a comparison of Google trends and Baidu index (2015) 0.01
    0.0053104796 = product of:
      0.015931439 = sum of:
        0.015931439 = product of:
          0.031862877 = sum of:
            0.031862877 = weight(_text_:22 in 1605) [ClassicSimilarity], result of:
              0.031862877 = score(doc=1605,freq=2.0), product of:
                0.1647081 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047034867 = queryNorm
                0.19345059 = fieldWeight in 1605, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1605)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Source
    Journal of the Association for Information Science and Technology. 66(2015) no.1, S.13-22
  19. Fonseca, F.; Marcinkowski, M.; Davis, C.: Cyber-human systems of thought and understanding (2019) 0.01
    0.0053104796 = product of:
      0.015931439 = sum of:
        0.015931439 = product of:
          0.031862877 = sum of:
            0.031862877 = weight(_text_:22 in 5011) [ClassicSimilarity], result of:
              0.031862877 = score(doc=5011,freq=2.0), product of:
                0.1647081 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047034867 = queryNorm
                0.19345059 = fieldWeight in 5011, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5011)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    7. 3.2019 16:32:22
  20. Peters, G.; Gaese, V.: ¬Das DocCat-System in der Textdokumentation von G+J (2003) 0.00
    0.0042483835 = product of:
      0.01274515 = sum of:
        0.01274515 = product of:
          0.0254903 = sum of:
            0.0254903 = weight(_text_:22 in 1507) [ClassicSimilarity], result of:
              0.0254903 = score(doc=1507,freq=2.0), product of:
                0.1647081 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047034867 = queryNorm
                0.15476047 = fieldWeight in 1507, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1507)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 4.2003 11:45:36

Years

Languages

  • e 18
  • d 7

Types