Search (9 results, page 1 of 1)

  • × author_ss:"Khoo, C."
  1. Poo, D.C.C.; Khoo, C.: Subject searching in online catalog systems (1997) 0.00
    0.00270615 = product of:
      0.0054123 = sum of:
        0.0054123 = product of:
          0.0108246 = sum of:
            0.0108246 = weight(_text_:a in 7760) [ClassicSimilarity], result of:
              0.0108246 = score(doc=7760,freq=2.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.20383182 = fieldWeight in 7760, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=7760)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  2. Lee, C.-H.; Khoo, C.; Na, J.-C.: Automatic identification of treatment relations for medical ontology learning : an exploratory study (2004) 0.00
    0.0023919214 = product of:
      0.0047838427 = sum of:
        0.0047838427 = product of:
          0.009567685 = sum of:
            0.009567685 = weight(_text_:a in 2661) [ClassicSimilarity], result of:
              0.009567685 = score(doc=2661,freq=16.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.18016359 = fieldWeight in 2661, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2661)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This study is part of a project to develop an automatic method to build ontologies, especially in a medical domain, from a document collection. An earlier study had investigated an approach to inferring semantic relations between medical concepts using the UMLS (Unified Medical Language System) semantic net. The study found that semantic relations between concepts could be inferred 68% of the time, although the method often could not distinguish between a few possible relation types. Our current research focuses an the use of natural language processing techniques to improve the identification of semantic relations. In particular, we explore both a semi-automatic and manual construction of linguistic patterns for identifying treatment relations in medical abstracts in the domain of colon cancer treatment. Association rule mining was applied to sample sentences containing both a disease concept and a reference to drug, to identify frequently occurring word pattems to see if these pattems could be used to identify treatment relations in sentences. This did not yield many useful patterns, suggesting that statistical association measures have to be complemented with syntactic and semantic constraints to identify useful patterns. In the second part of the study, linguistic patterns were manually constructed based an the same sentences. This yielded promising results. Work is ongoing to improve the manually constructed pattems as well as to identify the syntactic and semantic constraints that can be used to improve the automatic construction of linguistic patterns.
    Type
    a
  3. Wang, Z.; Chaudhry, A.S.; Khoo, C.: Support from bibliographic tools to build an organizational taxonomy for navigation : use of a general classification scheme and domain thesauri (2010) 0.00
    0.002269176 = product of:
      0.004538352 = sum of:
        0.004538352 = product of:
          0.009076704 = sum of:
            0.009076704 = weight(_text_:a in 3710) [ClassicSimilarity], result of:
              0.009076704 = score(doc=3710,freq=10.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.1709182 = fieldWeight in 3710, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3710)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    A study was conducted to investigate the capability of a general classification scheme and domain thesauri to support the construction of an organizational taxonomy to be used for navigation, and to develop steps and guidelines for constructing the hierarchical structure and categories. The study was conducted in the context of a graduate department in information studies in Singapore that offers Master's and PhD programs in information studies, information systems, and knowledge management. An organizational taxonomy, called Information Studies Taxonomy, was built for learning, teaching and research tasks of the department using the Dewey Decimal Classification and three domain thesauri (ASIS&T, LISA, and ERIC). The support and difficulties of using the general classification scheme and domain thesauri were identified in the taxonomy development process. Steps and guidelines for constructing the hierarchical structure and categories were developed based on problems encountered in using the sources.
    Type
    a
  4. Ou, S.; Khoo, C.; Goh, D.H.; Heng, H.-Y.: Automatic discourse parsing of sociology dissertation abstracts as sentence categorization (2004) 0.00
    0.0022438213 = product of:
      0.0044876426 = sum of:
        0.0044876426 = product of:
          0.008975285 = sum of:
            0.008975285 = weight(_text_:a in 2676) [ClassicSimilarity], result of:
              0.008975285 = score(doc=2676,freq=22.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.16900843 = fieldWeight in 2676, product of:
                  4.690416 = tf(freq=22.0), with freq of:
                    22.0 = termFreq=22.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2676)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    We investigated an approach to automatic discourse parsing of sociology dissertation abstracts as a sentence categorization task. Decision tree induction was used for the automatic categorization. Three models were developed. Model 1 made use of word tokens found in the sentences. Model 2 made use of both word tokens and sentence position in the abstract. In addition to the attributes used in Model 2, Model 3 also considered information regarding the presence of indicator words in surrounding sentences. Model 3 obtained the highest accuracy rate of 74.5 % when applied to a test sample, compared to 71.6% for Model 2 and 60.8% for Model 1. The results indicated that information about sentence position can substantially increase the accuracy of categorization, and indicator words in earlier sentences (before the sentence being processed) also contribute to the categorization accuracy.
    Content
    1. Introduction This paper reports our initial effort to develop an automatic method for parsing the discourse structure of sociology dissertation abstracts. This study is part of a broader study to develop a method for multi-document summarization. Accurate discourse parsing will make it easier to perform automatic multi-document summarization of dissertation abstracts. In a previous study, we determined that the macro-level structure of dissertation abstracts typically has five sections (Khoo et al., 2002). In this study, we treated discourse parsing as a text categorization problem - assigning each sentence in a dissertation abstract to one of the five predefined sections or categories. Decision tree induction, a machine-learning method, was applied to word tokens found in the abstracts to construct a decision tree model for the categorization purpose. Decision tree induction was selected primarily because decision tree models are easy to interpret and can be converted to rules that can be incorporated in other computer programs. A well-known decision-tree induction program, C5.0 (Quinlan, 1993), was used in this study.
    Type
    a
  5. Khoo, C.; Chan, S.; Niu, Y.: ¬The many facets of the cause-effect relation (2002) 0.00
    0.0020296127 = product of:
      0.0040592253 = sum of:
        0.0040592253 = product of:
          0.008118451 = sum of:
            0.008118451 = weight(_text_:a in 1192) [ClassicSimilarity], result of:
              0.008118451 = score(doc=1192,freq=8.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.15287387 = fieldWeight in 1192, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1192)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This chapter presents a broad survey of the cause-effect relation, with particular emphasis an how the relation is expressed in text. Philosophers have been grappling with the concept of causation for centuries. Researchers in social psychology have found that the human mind has a very complex mechanism for identifying and attributing the cause for an event. Inferring cause-effect relations between events and statements has also been found to be an important part of reading and text comprehension, especially for narrative text. Though many of the cause-effect relations in text are implied and have to be inferred by the reader, there is also a wide variety of linguistic expressions for explicitly indicating cause and effect. In addition, it has been found that certain words have "causal valence"-they bias the reader to attribute cause in certain ways. Cause-effect relations can also be divided into several different types.
    Type
    a
  6. Abdul, H.; Khoo, C.: Automatic indexing of medical literature using phrase matching : an exploratory study 0.00
    0.001913537 = product of:
      0.003827074 = sum of:
        0.003827074 = product of:
          0.007654148 = sum of:
            0.007654148 = weight(_text_:a in 3601) [ClassicSimilarity], result of:
              0.007654148 = score(doc=3601,freq=4.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.14413087 = fieldWeight in 3601, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3601)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Reports the 1st part of a study to apply the technique of phrase matching to the automatic assignment of MeSH subject headings and subheadings to abstracts of periodical articles.
    Type
    a
  7. Khoo, C.; Myaeng, S.H.: Identifying semantic relations in text for information retrieval and information extraction (2002) 0.00
    0.001757696 = product of:
      0.003515392 = sum of:
        0.003515392 = product of:
          0.007030784 = sum of:
            0.007030784 = weight(_text_:a in 1197) [ClassicSimilarity], result of:
              0.007030784 = score(doc=1197,freq=6.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.13239266 = fieldWeight in 1197, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1197)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Automatic identification of semantic relations in text is a difficult problem, but is important for many applications. It has been used for relation matching in information retrieval to retrieve documents that contain not only the concepts but also the relations between concepts specified in the user's query. It is an integral part of information extraction-extracting from natural language text, facts or pieces of information related to a particular event or topic. Other potential applications are in the construction of relational thesauri (semantic networks of related concepts) and other kinds of knowledge bases, and in natural language processing applications such as machine translation and computer comprehension of text. This chapter examines the main methods used for identifying semantic relations automatically and their application in information retrieval and information extraction.
    Type
    a
  8. Na, J.-C.; Sui, H.; Khoo, C.; Chan, S.; Zhou, Y.: Effectiveness of simple linguistic processing in automatic sentiment classification of product reviews (2004) 0.00
    0.0016913437 = product of:
      0.0033826875 = sum of:
        0.0033826875 = product of:
          0.006765375 = sum of:
            0.006765375 = weight(_text_:a in 2624) [ClassicSimilarity], result of:
              0.006765375 = score(doc=2624,freq=8.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.12739488 = fieldWeight in 2624, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2624)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This paper reports a study in automatic sentiment classification, i.e., automatically classifying documents as expressing positive or negative Sentiments/opinions. The study investigates the effectiveness of using SVM (Support Vector Machine) an various text features to classify product reviews into recommended (positive Sentiment) and not recommended (negative sentiment). Compared with traditional topical classification, it was hypothesized that syntactic and semantic processing of text would be more important for sentiment classification. In the first part of this study, several different approaches, unigrams (individual words), selected words (such as verb, adjective, and adverb), and words labelled with part-of-speech tags were investigated. A sample of 1,800 various product reviews was retrieved from Review Centre (www.reviewcentre.com) for the study. 1,200 reviews were used for training, and 600 for testing. Using SVM, the baseline unigram approach obtained an accuracy rate of around 76%. The use of selected words obtained a marginally better result of 77.33%. Error analysis suggests various approaches for improving classification accuracy: use of negation phrase, making inference from superficial words, and solving the problem of comments an parts. The second part of the study that is in progress investigates the use of negation phrase through simple linguistic processing to improve classification accuracy. This approach increased the accuracy rate up to 79.33%.
    Type
    a
  9. Zhonghong, W.; Chaudhry, A.S.; Khoo, C.: Potential and prospects of taxonomies for content organization (2006) 0.00
    0.001674345 = product of:
      0.00334869 = sum of:
        0.00334869 = product of:
          0.00669738 = sum of:
            0.00669738 = weight(_text_:a in 169) [ClassicSimilarity], result of:
              0.00669738 = score(doc=169,freq=4.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.12611452 = fieldWeight in 169, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=169)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    While taxonomies are being increasingly discussed in published and grey literature, the term taxonomy still seems to be stated quite loosely and obscurely. This paper aims at explaining and clarifying the concept of taxonomy in the context of information organization. To this end, the salient features of taxonomies are identified and their scope, nature, and role are further elaborated based on an extensive literature review. In the meantime, the connection and distinctions between taxonomies and classification schemes and thesauri are also identified, and the rationale that taxonomies are chosen as a viable knowledge organization system used in organization-wide websites to support browsing and aid navigation is clarified.
    Type
    a