Search (3 results, page 1 of 1)

  • × author_ss:"Ke, H.-R."
  • × language_ss:"e"
  1. Yeh, J.-Y.; Ke, H.-R.; Yang, W.-P.; Meng, I.-H.: Text summarization using a trainable summarizer and latent semantic analysis (2005) 0.00
    0.0023919214 = product of:
      0.0047838427 = sum of:
        0.0047838427 = product of:
          0.009567685 = sum of:
            0.009567685 = weight(_text_:a in 1003) [ClassicSimilarity], result of:
              0.009567685 = score(doc=1003,freq=16.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.18016359 = fieldWeight in 1003, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1003)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This paper proposes two approaches to address text summarization: modified corpus-based approach (MCBA) and LSA-based T.R.M. approach (LSA + T.R.M.). The first is a trainable summarizer, which takes into account several features, including position, positive keyword, negative keyword, centrality, and the resemblance to the title, to generate summaries. Two new ideas are exploited: (1) sentence positions are ranked to emphasize the significances of different sentence positions, and (2) the score function is trained by the genetic algorithm (GA) to obtain a suitable combination of feature weights. The second uses latent semantic analysis (LSA) to derive the semantic matrix of a document or a corpus and uses semantic sentence representation to construct a semantic text relationship map. We evaluate LSA + T.R.M. both with single documents and at the corpus level to investigate the competence of LSA in text summarization. The two novel approaches were measured at several compression rates on a data corpus composed of 100 political articles. When the compression rate was 30%, an average f-measure of 49% for MCBA, 52% for MCBA + GA, 44% and 40% for LSA + T.R.M. in single-document and corpus level were achieved respectively.
    Type
    a
  2. Chen, Y.-N.; Ke, H.-R.: ¬A study on mental models of taggers and experts for article indexing based on analysis of keyword usage (2014) 0.00
    0.0018909799 = product of:
      0.0037819599 = sum of:
        0.0037819599 = product of:
          0.0075639198 = sum of:
            0.0075639198 = weight(_text_:a in 1334) [ClassicSimilarity], result of:
              0.0075639198 = score(doc=1334,freq=10.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.14243183 = fieldWeight in 1334, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1334)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This article explores the mental models of article indexing of taggers and experts in keyword usage. Better understanding of the mental models of taggers and experts and their usage gap may inspire better selection of appropriate keywords for organizing information resources. Using a data set of 3,972 tags from CiteULike and 6,708 descriptors from Library and Information Science Abstracts (LISA) from 1,489 scholarly articles of 13 library and information science journals, social network analysis and frequent-pattern tree methods were used to capture and build up the mental models of article indexing of taggers and experts when using keywords, and to generalize their structures and patterns. When measured with respect to the terms used, a power-law distribution, a comparison of terms used as tags and descriptors, social network analysis (including centrality, overall structure, and role equivalent) and frequent-pattern tree analysis, little similarity was found between the mental models of taggers and experts. Twenty-five patterns of path-based rules and 12 identical rules of frequent-pattern trees were shared by taggers and experts. Title- and topic-related keyword categories were the most popular keyword categories used in path-based rules of frequent-pattern trees, and also the most popular members of 25 patterns and the starting point of the 12 identical rules.
    Type
    a
  3. Huang, S.-H.; Ke, H.-R.; Yang, W.-P.: Enhancing semantic digital library query using a content and service inference model (CSIM) (2005) 0.00
    0.001757696 = product of:
      0.003515392 = sum of:
        0.003515392 = product of:
          0.007030784 = sum of:
            0.007030784 = weight(_text_:a in 1054) [ClassicSimilarity], result of:
              0.007030784 = score(doc=1054,freq=6.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.13239266 = fieldWeight in 1054, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1054)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Although digital library (DL) information is becoming increasingly annotated using metadata, semantic query with respect to the structure of metadata has seldom been addressed. The correlation of the two important aspects of DL--content and services--can generate additional semantic relationships. This study proposes a content and service inference model (CSIM) to derive 15 relationships between content and services, and defines functions to manipulate these relationships. Adding the manipulation functions to query predicates facilitates the description of structural semantics of DL content. Moreover, in search for DL services, inferences concerning CSIM relationships can be made to reuse DL service components. Highly promising with experimental results demonstrates that CSIM outperforms the conventional keyword-based method in both content and service queries. Applying CSIM in DL significantly improves semantic queries and alleviates the administrative load when developing novel DL services such as DL query interface, library resource-planning and virtual union catalog system.
    Type
    a