Search (1176 results, page 1 of 59)

  • × year_i:[1990 TO 2000}
  • × type_ss:"a"
  1. Chowdhury, G.G.: Template mining for information extraction from digital documents (1999) 0.22
    0.22446406 = product of:
      0.44892812 = sum of:
        0.44892812 = sum of:
          0.35283166 = weight(_text_:mining in 4577) [ClassicSimilarity], result of:
            0.35283166 = score(doc=4577,freq=4.0), product of:
              0.28585905 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.05066224 = queryNorm
              1.2342855 = fieldWeight in 4577, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.109375 = fieldNorm(doc=4577)
          0.09609647 = weight(_text_:22 in 4577) [ClassicSimilarity], result of:
            0.09609647 = score(doc=4577,freq=2.0), product of:
              0.17741053 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05066224 = queryNorm
              0.5416616 = fieldWeight in 4577, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.109375 = fieldNorm(doc=4577)
      0.5 = coord(1/2)
    
    Date
    2. 4.2000 18:01:22
    Theme
    Data Mining
  2. Matson, L.D.; Bonski, D.J.: Do digital libraries need librarians? (1997) 0.13
    0.12826519 = product of:
      0.25653037 = sum of:
        0.25653037 = sum of:
          0.2016181 = weight(_text_:mining in 1737) [ClassicSimilarity], result of:
            0.2016181 = score(doc=1737,freq=4.0), product of:
              0.28585905 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.05066224 = queryNorm
              0.705306 = fieldWeight in 1737, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.0625 = fieldNorm(doc=1737)
          0.054912273 = weight(_text_:22 in 1737) [ClassicSimilarity], result of:
            0.054912273 = score(doc=1737,freq=2.0), product of:
              0.17741053 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05066224 = queryNorm
              0.30952093 = fieldWeight in 1737, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=1737)
      0.5 = coord(1/2)
    
    Abstract
    Defines digital libraries and discusses the effects of new technology on librarians. Examines the different viewpoints of librarians and information technologists on digital libraries. Describes the development of a digital library at the National Drug Intelligence Center, USA, which was carried out in collaboration with information technology experts. The system is based on Web enabled search technology to find information, data visualization and data mining to visualize it and use of SGML as an information standard to store it
    Date
    22.11.1998 18:57:22
    Theme
    Data Mining
  3. Saz, J.T.: Perspectivas en recuperacion y explotacion de informacion electronica : el 'data mining' (1997) 0.11
    0.109129 = product of:
      0.218258 = sum of:
        0.218258 = product of:
          0.436516 = sum of:
            0.436516 = weight(_text_:mining in 3723) [ClassicSimilarity], result of:
              0.436516 = score(doc=3723,freq=12.0), product of:
                0.28585905 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.05066224 = queryNorm
                1.5270323 = fieldWeight in 3723, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3723)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Presents the concept and the techniques identified by the term data mining. Explains the principles and phases of developing a data mining process, and the main types of data mining tools
    Footnote
    Übers. des Titels: Perspectives on the retrieval and exploitation of electronic information: data mining
    Theme
    Data Mining
  4. Tunbridge, N.: Semiology put to data mining (1999) 0.10
    0.10080905 = product of:
      0.2016181 = sum of:
        0.2016181 = product of:
          0.4032362 = sum of:
            0.4032362 = weight(_text_:mining in 6782) [ClassicSimilarity], result of:
              0.4032362 = score(doc=6782,freq=4.0), product of:
                0.28585905 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.05066224 = queryNorm
                1.410612 = fieldWeight in 6782, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.125 = fieldNorm(doc=6782)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Theme
    Data Mining
  5. Spertus, E.: ParaSite : mining structural information on the Web (1997) 0.10
    0.098738894 = product of:
      0.19747779 = sum of:
        0.19747779 = sum of:
          0.14256552 = weight(_text_:mining in 2740) [ClassicSimilarity], result of:
            0.14256552 = score(doc=2740,freq=2.0), product of:
              0.28585905 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.05066224 = queryNorm
              0.49872664 = fieldWeight in 2740, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.0625 = fieldNorm(doc=2740)
          0.054912273 = weight(_text_:22 in 2740) [ClassicSimilarity], result of:
            0.054912273 = score(doc=2740,freq=2.0), product of:
              0.17741053 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05066224 = queryNorm
              0.30952093 = fieldWeight in 2740, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=2740)
      0.5 = coord(1/2)
    
    Date
    1. 8.1996 22:08:06
  6. Amir, A.; Feldman, R.; Kashi, R.: ¬A new and versatile method for association generation (1997) 0.10
    0.098738894 = product of:
      0.19747779 = sum of:
        0.19747779 = sum of:
          0.14256552 = weight(_text_:mining in 1270) [ClassicSimilarity], result of:
            0.14256552 = score(doc=1270,freq=2.0), product of:
              0.28585905 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.05066224 = queryNorm
              0.49872664 = fieldWeight in 1270, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.0625 = fieldNorm(doc=1270)
          0.054912273 = weight(_text_:22 in 1270) [ClassicSimilarity], result of:
            0.054912273 = score(doc=1270,freq=2.0), product of:
              0.17741053 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05066224 = queryNorm
              0.30952093 = fieldWeight in 1270, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=1270)
      0.5 = coord(1/2)
    
    Source
    Information systems. 22(1997) nos.5/6, S.333-347
    Theme
    Data Mining
  7. Lawson, M.: Automatic extraction of citations from the text of English-language patents : an example of template mining (1996) 0.10
    0.09619889 = product of:
      0.19239777 = sum of:
        0.19239777 = sum of:
          0.15121357 = weight(_text_:mining in 2654) [ClassicSimilarity], result of:
            0.15121357 = score(doc=2654,freq=4.0), product of:
              0.28585905 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.05066224 = queryNorm
              0.5289795 = fieldWeight in 2654, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.046875 = fieldNorm(doc=2654)
          0.0411842 = weight(_text_:22 in 2654) [ClassicSimilarity], result of:
            0.0411842 = score(doc=2654,freq=2.0), product of:
              0.17741053 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05066224 = queryNorm
              0.23214069 = fieldWeight in 2654, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2654)
      0.5 = coord(1/2)
    
    Abstract
    Describes and evaluates methods for automatically isolating and extracting biliographic references from the full texts of patents, designed to facilitate the work of patent examiners who currently perform this task manually. These references include citations both to patents and to other bibliographic sources. Notes that patents are unusual as citing documents in that the citations occur maily in the body of the text, rather than as footnotes or in separate sections. Describes the natural language processing technique of template mining used to extract data directly from the text where either the data or the text surrounding the data form recognizable patterns. When text matches a template, the system extracts data according to instructions associated with that template. Examines the sub languages of citations and the development of templates for the extraction of citations to patent. Reports results of running 2 reference extraction systems against a sample of 100 European Patent Office patent documents, with recall and prescision data for patent and non patent citations, and concludes with suggestions for future improvements
    Source
    Journal of information science. 22(1996) no.6, S.423-436
  8. Li, D.: Knowledge representation and discovery based on linguistic atoms (1998) 0.10
    0.09619889 = product of:
      0.19239777 = sum of:
        0.19239777 = sum of:
          0.15121357 = weight(_text_:mining in 3836) [ClassicSimilarity], result of:
            0.15121357 = score(doc=3836,freq=4.0), product of:
              0.28585905 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.05066224 = queryNorm
              0.5289795 = fieldWeight in 3836, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.046875 = fieldNorm(doc=3836)
          0.0411842 = weight(_text_:22 in 3836) [ClassicSimilarity], result of:
            0.0411842 = score(doc=3836,freq=2.0), product of:
              0.17741053 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05066224 = queryNorm
              0.23214069 = fieldWeight in 3836, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=3836)
      0.5 = coord(1/2)
    
    Abstract
    Describes a new concept of linguistic atoms with 3 digital characteristics: expected value Ex, entropy En, and deviation D. The mathematical description has effectively integrated the fuzziness and randomness of linguistic terms in a unified way. Develops a method of knowledge representation in KDD, which bridges the gap between quantitative and qualitative knowledge. Mapping between quantities and qualities becomes much easier and interchangeable. In order to discover generalised knowledge from a database, uses virtual linguistic terms and cloud transfer for the auto-generation of concept hierarchies to attributes. Predicitve data mining with the cloud model is given for implementation. Illustrates the advantages of this linguistic model in KDD
    Footnote
    Contribution to a special issue of selected papers from the Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD'97), held Singapore, 22-23 Feb 1997
  9. Fayyad, U.; Piatetsky-Shapiro, G.; Smyth, P.: From data mining to knowledge discovery in databases (1996) 0.09
    0.08910345 = product of:
      0.1782069 = sum of:
        0.1782069 = product of:
          0.3564138 = sum of:
            0.3564138 = weight(_text_:mining in 7458) [ClassicSimilarity], result of:
              0.3564138 = score(doc=7458,freq=8.0), product of:
                0.28585905 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.05066224 = queryNorm
                1.2468166 = fieldWeight in 7458, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.078125 = fieldNorm(doc=7458)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Gives an overview of data mining and knowledge discovery in databases. Clarifies how they are related both to each other and to related fields. Mentions real world applications data mining techniques, challenges involved in real world applications of knowledge discovery, and current and future research directions
    Theme
    Data Mining
  10. Schmid, J.: Data mining : wie finde ich in Datensammlungen entscheidungsrelevante Muster? (1999) 0.09
    0.088207915 = product of:
      0.17641583 = sum of:
        0.17641583 = product of:
          0.35283166 = sum of:
            0.35283166 = weight(_text_:mining in 4540) [ClassicSimilarity], result of:
              0.35283166 = score(doc=4540,freq=4.0), product of:
                0.28585905 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.05066224 = queryNorm
                1.2342855 = fieldWeight in 4540, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4540)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Theme
    Data Mining
  11. Koczkodaj, W.: ¬A note on using a consistency-driven approach to CD-ROM selection (1997) 0.09
    0.08639653 = product of:
      0.17279306 = sum of:
        0.17279306 = sum of:
          0.12474483 = weight(_text_:mining in 7893) [ClassicSimilarity], result of:
            0.12474483 = score(doc=7893,freq=2.0), product of:
              0.28585905 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.05066224 = queryNorm
              0.4363858 = fieldWeight in 7893, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.0546875 = fieldNorm(doc=7893)
          0.048048235 = weight(_text_:22 in 7893) [ClassicSimilarity], result of:
            0.048048235 = score(doc=7893,freq=2.0), product of:
              0.17741053 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05066224 = queryNorm
              0.2708308 = fieldWeight in 7893, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=7893)
      0.5 = coord(1/2)
    
    Abstract
    As with print collections, the evaluation and selection of CD-ROMs should be based on established guidelines. Such attributes as computer network compatibility and platform are exclusively applicable to CD-ROM. Presents a knowledge based system to prioritize and select CD-ROMs for a library collection, operating on consistency driven pairwise comparisons. The computer system indicates the most inconsistent judgements and allows librarians to reconsider their position. After consistency analysis is completed, the software computes the weights of all criteria used in the evaluation process. The system includes a subsystem for evaluating CD-ROM titles. Offers a CD-ROM evaluation form. Discusses cost considerations; the use of pairwise comparisons in knowledge based systems with reference to data mining; the CD-ROM selection process; and consistency analysis of experts' judgements
    Date
    6. 3.1997 16:22:15
  12. Hofstede, A.H.M. ter; Proper, H.A.; Van der Weide, T.P.: Exploiting fact verbalisation in conceptual information modelling (1997) 0.09
    0.08639653 = product of:
      0.17279306 = sum of:
        0.17279306 = sum of:
          0.12474483 = weight(_text_:mining in 2908) [ClassicSimilarity], result of:
            0.12474483 = score(doc=2908,freq=2.0), product of:
              0.28585905 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.05066224 = queryNorm
              0.4363858 = fieldWeight in 2908, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2908)
          0.048048235 = weight(_text_:22 in 2908) [ClassicSimilarity], result of:
            0.048048235 = score(doc=2908,freq=2.0), product of:
              0.17741053 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05066224 = queryNorm
              0.2708308 = fieldWeight in 2908, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2908)
      0.5 = coord(1/2)
    
    Source
    Information systems. 22(1997) nos.5/6, S.349-385
    Theme
    Data Mining
  13. Cheung, D.W.; Kao, B.; Lee, J.: Discovering user access patterns on the World Wide Web (1998) 0.09
    0.08639653 = product of:
      0.17279306 = sum of:
        0.17279306 = sum of:
          0.12474483 = weight(_text_:mining in 332) [ClassicSimilarity], result of:
            0.12474483 = score(doc=332,freq=2.0), product of:
              0.28585905 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.05066224 = queryNorm
              0.4363858 = fieldWeight in 332, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.0546875 = fieldNorm(doc=332)
          0.048048235 = weight(_text_:22 in 332) [ClassicSimilarity], result of:
            0.048048235 = score(doc=332,freq=2.0), product of:
              0.17741053 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05066224 = queryNorm
              0.2708308 = fieldWeight in 332, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=332)
      0.5 = coord(1/2)
    
    Footnote
    Contribution to a special issue of selected papers from the Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD'97), held Singapore, 22-23 Feb 1997
  14. Raghavan, V.V.; Deogun, J.S.; Sever, H.: Knowledge discovery and data mining : introduction (1998) 0.08
    0.082510956 = product of:
      0.16502191 = sum of:
        0.16502191 = product of:
          0.33004382 = sum of:
            0.33004382 = weight(_text_:mining in 2899) [ClassicSimilarity], result of:
              0.33004382 = score(doc=2899,freq=14.0), product of:
                0.28585905 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.05066224 = queryNorm
                1.1545684 = fieldWeight in 2899, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2899)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Defines knowledge discovery and database mining. The challenge for knowledge discovery in databases (KDD) is to automatically process large quantities of raw data, identifying the most significant and meaningful patterns, and present these as as knowledge appropriate for achieving a user's goals. Data mining is the process of deriving useful knowledge from real world databases through the application of pattern extraction techniques. Explains the goals of, and motivation for, research work on data mining. Discusses the nature of database contents, along with problems within the field of data mining
    Footnote
    Contribution to a special issue devoted to knowledge discovery and data mining
    Theme
    Data Mining
  15. Fayyad, U.M.: Data mining and knowledge dicovery : making sense out of data (1996) 0.08
    0.077165864 = product of:
      0.15433173 = sum of:
        0.15433173 = product of:
          0.30866346 = sum of:
            0.30866346 = weight(_text_:mining in 7007) [ClassicSimilarity], result of:
              0.30866346 = score(doc=7007,freq=6.0), product of:
                0.28585905 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.05066224 = queryNorm
                1.079775 = fieldWeight in 7007, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.078125 = fieldNorm(doc=7007)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Defines knowledge discovery and data mining (KDD) as the overall process of extracting high level knowledge from low level data. Outlines the KDD process. Explains how KDD is related to the fields of: statistics, pattern recognition, machine learning, artificial intelligence, databases and data warehouses
    Theme
    Data Mining
  16. Priss, U.: Description logic and faceted knowledge representation (1999) 0.07
    0.074054174 = product of:
      0.14810835 = sum of:
        0.14810835 = sum of:
          0.10692415 = weight(_text_:mining in 2655) [ClassicSimilarity], result of:
            0.10692415 = score(doc=2655,freq=2.0), product of:
              0.28585905 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.05066224 = queryNorm
              0.37404498 = fieldWeight in 2655, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.046875 = fieldNorm(doc=2655)
          0.0411842 = weight(_text_:22 in 2655) [ClassicSimilarity], result of:
            0.0411842 = score(doc=2655,freq=2.0), product of:
              0.17741053 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05066224 = queryNorm
              0.23214069 = fieldWeight in 2655, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2655)
      0.5 = coord(1/2)
    
    Abstract
    The term "facet" was introduced into the field of library classification systems by Ranganathan in the 1930's [Ranganathan, 1962]. A facet is a viewpoint or aspect. In contrast to traditional classification systems, faceted systems are modular in that a domain is analyzed in terms of baseline facets which are then synthesized. In this paper, the term "facet" is used in a broader meaning. Facets can describe different aspects on the same level of abstraction or the same aspect on different levels of abstraction. The notion of facets is related to database views, multicontexts and conceptual scaling in formal concept analysis [Ganter and Wille, 1999], polymorphism in object-oriented design, aspect-oriented programming, views and contexts in description logic and semantic networks. This paper presents a definition of facets in terms of faceted knowledge representation that incorporates the traditional narrower notion of facets and potentially facilitates translation between different knowledge representation formalisms. A goal of this approach is a modular, machine-aided knowledge base design mechanism. A possible application is faceted thesaurus construction for information retrieval and data mining. Reasoning complexity depends on the size of the modules (facets). A more general analysis of complexity will be left for future research.
    Date
    22. 1.2016 17:30:31
  17. Howlett, D.: Digging deep for treasure (1998) 0.07
    0.07128276 = product of:
      0.14256552 = sum of:
        0.14256552 = product of:
          0.28513104 = sum of:
            0.28513104 = weight(_text_:mining in 4544) [ClassicSimilarity], result of:
              0.28513104 = score(doc=4544,freq=2.0), product of:
                0.28585905 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.05066224 = queryNorm
                0.9974533 = fieldWeight in 4544, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.125 = fieldNorm(doc=4544)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Theme
    Data Mining
  18. Lingras, P.J.; Yao, Y.Y.: Data mining using extensions of the rough set model (1998) 0.07
    0.069734484 = product of:
      0.13946897 = sum of:
        0.13946897 = product of:
          0.27893794 = sum of:
            0.27893794 = weight(_text_:mining in 2910) [ClassicSimilarity], result of:
              0.27893794 = score(doc=2910,freq=10.0), product of:
                0.28585905 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.05066224 = queryNorm
                0.97578835 = fieldWeight in 2910, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2910)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Examines basic issues of data mining using the theory of rough sets, which is a recent proposal for generalizing classical set theory. The Pawlak rough set model is based on the concept of an equivalence relation. A generalized rough set model need not be based on equivalence relation axioms. The Pawlak rough set model has been used for deriving deterministic as well as probabilistic rules froma complete database. Demonstrates that a generalised rough set model can be used for generating rules from incomplete databases. These rules are based on plausability functions proposed by Shafer. Discusses the importance of rule extraction from incomplete databases in data mining
    Footnote
    Contribution to a special issue devoted to knowledge discovery and data mining
    Theme
    Data Mining
  19. Trybula, W.J.: Data mining and knowledge discovery (1997) 0.06
    0.062372416 = product of:
      0.12474483 = sum of:
        0.12474483 = product of:
          0.24948967 = sum of:
            0.24948967 = weight(_text_:mining in 2300) [ClassicSimilarity], result of:
              0.24948967 = score(doc=2300,freq=8.0), product of:
                0.28585905 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.05066224 = queryNorm
                0.8727716 = fieldWeight in 2300, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2300)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    State of the art review of the recently developed concepts of data mining (defined as the automated process of evaluating data and finding relationships) and knowledge discovery (defined as the automated process of extracting information, especially unpredicted relationships or previously unknown patterns among the data) with particular reference to numerical data. Includes: the knowledge acquisition process; data mining; evaluation methods; and knowledge discovery. Concludes that existing work in the field are confusing because the terminology is inconsistent and poorly defined. Although methods are available for analyzing and cleaning databases, better coordinated efforts should be directed toward providing users with improved means of structuring search mechanisms to explore the data for relationships
    Theme
    Data Mining
  20. Wu, X.: Rule induction with extension matrices (1998) 0.05
    0.053462073 = product of:
      0.10692415 = sum of:
        0.10692415 = product of:
          0.2138483 = sum of:
            0.2138483 = weight(_text_:mining in 2912) [ClassicSimilarity], result of:
              0.2138483 = score(doc=2912,freq=8.0), product of:
                0.28585905 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.05066224 = queryNorm
                0.74808997 = fieldWeight in 2912, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2912)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Presents a heuristic, attribute-based, noise-tolerant data mining program, HCV (Version 2.0), absed on the newly-developed extension matrix approach. Gives a simple example of attribute-based induction to show the difference between the rules in variable-valued logic produced by HCV, the decision tree generated by C4.5 and the decision tree's decompiled rules by C4.5 rules. Outlines the extension matrix approach for data mining. Describes the HCV algorithm in detail. Outlines techniques developed and implemented in the HCV program for noise handling and discretization of continuous domains respectively. Follows these with a performance comparison of HCV with famous ID3-like algorithms including C4.5 and C4.5 rules on a collection of standard databases including the famous MONK's problems
    Footnote
    Contribution to a special issue devoted to knowledge discovery and data mining
    Theme
    Data Mining

Languages

Types

Themes