Search (5 results, page 1 of 1)

  • × author_ss:"Wong, S.K.M."
  1. Wong, S.K.M.: On modelling information retrieval with probabilistic inference (1995) 0.01
    0.0073227976 = product of:
      0.05125958 = sum of:
        0.011574914 = weight(_text_:information in 1938) [ClassicSimilarity], result of:
          0.011574914 = score(doc=1938,freq=6.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.2687516 = fieldWeight in 1938, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=1938)
        0.03968467 = weight(_text_:retrieval in 1938) [ClassicSimilarity], result of:
          0.03968467 = score(doc=1938,freq=8.0), product of:
            0.07421378 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.02453417 = queryNorm
            0.5347345 = fieldWeight in 1938, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=1938)
      0.14285715 = coord(2/14)
    
    Abstract
    Examines and extends the logical models of information retrieval in the context of probability theory and extends the applications of these fundamental ideas to term weighting and relevance. Develops a unified framework for modelling the retrieval process with probabilistic inference to provide a common conceptual and mathematical basis for many retrieval models, such as Boolean, fuzzy sets, vector space, and conventional probabilistic models. Employs this framework to identify the underlying assumptions by each model and analyzes the inherent relationships between them. Although the treatment is primarily theoretical, practical methods for rstimating the required probabilities are provided by simple examples
    Source
    ACM transactions on information systems. 13(1995) no.1, S.38-68
  2. Wong, S.K.M.; Yao, Y.Y.: Query formulation in linear retrieval models (1990) 0.00
    0.0037893022 = product of:
      0.026525114 = sum of:
        0.006682779 = weight(_text_:information in 3571) [ClassicSimilarity], result of:
          0.006682779 = score(doc=3571,freq=2.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.1551638 = fieldWeight in 3571, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=3571)
        0.019842334 = weight(_text_:retrieval in 3571) [ClassicSimilarity], result of:
          0.019842334 = score(doc=3571,freq=2.0), product of:
            0.07421378 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.02453417 = queryNorm
            0.26736724 = fieldWeight in 3571, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=3571)
      0.14285715 = coord(2/14)
    
    Source
    Journal of the American Society for Information Science. 41(1990) no.5, S.334-341
  3. Wong, S.K.M.; Butz, C.J.; Xiang, X.: Automated database schema design using mined data dependencies (1998) 0.00
    0.0035242445 = product of:
      0.02466971 = sum of:
        0.018822279 = weight(_text_:system in 2897) [ClassicSimilarity], result of:
          0.018822279 = score(doc=2897,freq=2.0), product of:
            0.07727166 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.02453417 = queryNorm
            0.2435858 = fieldWeight in 2897, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2897)
        0.0058474317 = weight(_text_:information in 2897) [ClassicSimilarity], result of:
          0.0058474317 = score(doc=2897,freq=2.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.13576832 = fieldWeight in 2897, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2897)
      0.14285715 = coord(2/14)
    
    Abstract
    Data dependencies are used in database schema design to enforce the correctness of a database as well as to reduce redundant data. These dependencies are usually determined from the semantics of the attributes and are then enforced upon the relations. Describes a bottom-up procedure for discovering multivalued dependencies in observed data without knowing a priori the relationships among the attributes. The proposed algorithm is an application of the technique designed for learning conditional independencies in probabilistic reasoning. A prototype system for automated database schema design has been implemented. Experiments were carried out to demonstrate both the effectiveness and efficiency of the method
    Source
    Journal of the American Society for Information Science. 49(1998) no.5, S.455-470
  4. Wong, S.K.M.; Yao, Y.Y.: ¬An information-theoretic measure of term specifics (1992) 0.00
    8.353474E-4 = product of:
      0.011694863 = sum of:
        0.011694863 = weight(_text_:information in 4807) [ClassicSimilarity], result of:
          0.011694863 = score(doc=4807,freq=8.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.27153665 = fieldWeight in 4807, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4807)
      0.071428575 = coord(1/14)
    
    Abstract
    The inverse document frequency (IDF) and signal-noise ratio (S/N) approaches are term weighting schemes based on term specifics. However, the existing justifications for these methods are still some what inconclusive and sometimes even based on incompatible assumptions. Introduces an information-theoretic measure for term specifics. Shows that the IDF weighting scheme can be derived from the proposed approach by assuming that the frequency of occurrence of each index term is uniform within the set of documents containing the term. The information-theoretic interpretation of term specifics also establishes the relationship between the IDF and S/N methods
    Source
    Journal of the American Society for Information Science. 43(1992) no.1, S.54-61
  5. Wong, S.K.M.; Yao, Y.Y.; Salton, G.; Buckley, C.: Evaluation of an adaptive linear model (1991) 0.00
    4.7734138E-4 = product of:
      0.006682779 = sum of:
        0.006682779 = weight(_text_:information in 4836) [ClassicSimilarity], result of:
          0.006682779 = score(doc=4836,freq=2.0), product of:
            0.04306919 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02453417 = queryNorm
            0.1551638 = fieldWeight in 4836, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=4836)
      0.071428575 = coord(1/14)
    
    Source
    Journal of the American Society for Information Science. 42(1991) no.10, S.723-730