Search (6 results, page 1 of 1)

  • × author_ss:"Chen, H.-H."
  1. Lee, Y.-Y.; Ke, H.; Yen, T.-Y.; Huang, H.-H.; Chen, H.-H.: Combining and learning word embedding with WordNet for semantic relatedness and similarity measurement (2020) 0.02
    0.01889613 = product of:
      0.09825988 = sum of:
        0.015683282 = weight(_text_:23 in 5871) [ClassicSimilarity], result of:
          0.015683282 = score(doc=5871,freq=2.0), product of:
            0.06600935 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.018417481 = queryNorm
            0.23759183 = fieldWeight in 5871, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=5871)
        0.015683282 = weight(_text_:23 in 5871) [ClassicSimilarity], result of:
          0.015683282 = score(doc=5871,freq=2.0), product of:
            0.06600935 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.018417481 = queryNorm
            0.23759183 = fieldWeight in 5871, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=5871)
        0.015683282 = weight(_text_:23 in 5871) [ClassicSimilarity], result of:
          0.015683282 = score(doc=5871,freq=2.0), product of:
            0.06600935 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.018417481 = queryNorm
            0.23759183 = fieldWeight in 5871, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=5871)
        0.015683282 = weight(_text_:23 in 5871) [ClassicSimilarity], result of:
          0.015683282 = score(doc=5871,freq=2.0), product of:
            0.06600935 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.018417481 = queryNorm
            0.23759183 = fieldWeight in 5871, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=5871)
        0.035526752 = weight(_text_:art in 5871) [ClassicSimilarity], result of:
          0.035526752 = score(doc=5871,freq=4.0), product of:
            0.08354246 = queryWeight, product of:
              4.5360413 = idf(docFreq=1287, maxDocs=44218)
              0.018417481 = queryNorm
            0.42525387 = fieldWeight in 5871, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.5360413 = idf(docFreq=1287, maxDocs=44218)
              0.046875 = fieldNorm(doc=5871)
      0.1923077 = coord(5/26)
    
    Abstract
    In this research, we propose 3 different approaches to measure the semantic relatedness between 2 words: (i) boost the performance of GloVe word embedding model via removing or transforming abnormal dimensions; (ii) linearly combine the information extracted from WordNet and word embeddings; and (iii) utilize word embedding and 12 linguistic information extracted from WordNet as features for Support Vector Regression. We conducted our experiments on 8 benchmark data sets, and computed Spearman correlations between the outputs of our methods and the ground truth. We report our results together with 3 state-of-the-art approaches. The experimental results show that our method can outperform state-of-the-art approaches in all the selected English benchmark data sets.
    Date
    23. 5.2020 16:19:47
  2. Chen, H.-H.; Lin, W.-C.; Yang, C.; Lin, W.-H.: Translating-transliterating named entities for multilingual information access (2006) 0.00
    0.0013808953 = product of:
      0.017951638 = sum of:
        0.012129237 = weight(_text_:5 in 1080) [ClassicSimilarity], result of:
          0.012129237 = score(doc=1080,freq=2.0), product of:
            0.05374404 = queryWeight, product of:
              2.9180994 = idf(docFreq=6494, maxDocs=44218)
              0.018417481 = queryNorm
            0.22568524 = fieldWeight in 1080, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.9180994 = idf(docFreq=6494, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1080)
        0.0058224006 = product of:
          0.0174672 = sum of:
            0.0174672 = weight(_text_:22 in 1080) [ClassicSimilarity], result of:
              0.0174672 = score(doc=1080,freq=2.0), product of:
                0.06449488 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.018417481 = queryNorm
                0.2708308 = fieldWeight in 1080, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1080)
          0.33333334 = coord(1/3)
      0.07692308 = coord(2/26)
    
    Date
    4. 6.2006 19:52:22
    Source
    Journal of the American Society for Information Science and Technology. 57(2006) no.5, S.645-659
  3. Tsai, M.-.F.; Chen, H.-H.; Wang, Y.-T.: Learning a merge model for multilingual information retrieval (2011) 0.00
    4.7124538E-4 = product of:
      0.012252379 = sum of:
        0.012252379 = weight(_text_:5 in 2750) [ClassicSimilarity], result of:
          0.012252379 = score(doc=2750,freq=4.0), product of:
            0.05374404 = queryWeight, product of:
              2.9180994 = idf(docFreq=6494, maxDocs=44218)
              0.018417481 = queryNorm
            0.22797652 = fieldWeight in 2750, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.9180994 = idf(docFreq=6494, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2750)
      0.03846154 = coord(1/26)
    
    Abstract
    This paper proposes a learning approach for the merging process in multilingual information retrieval (MLIR). To conduct the learning approach, we present a number of features that may influence the MLIR merging process. These features are mainly extracted from three levels: query, document, and translation. After the feature extraction, we then use the FRank ranking algorithm to construct a merge model. To the best of our knowledge, this practice is the first attempt to use a learning-based ranking algorithm to construct a merge model for MLIR merging. In our experiments, three test collections for the task of crosslingual information retrieval (CLIR) in NTCIR3, 4, and 5 are employed to assess the performance of our proposed method. Moreover, several merging methods are also carried out for a comparison, including traditional merging methods, the 2-step merging strategy, and the merging method based on logistic regression. The experimental results show that our proposed method can significantly improve merging quality on two different types of datasets. In addition to the effectiveness, through the merge model generated by FRank, our method can further identify key factors that influence the merging process. This information might provide us more insight and understanding into MLIR merging.
    Source
    Information processing and management. 47(2011) no.5, S.635-646
  4. Lee, L.-H.; Juan, Y.-C.; Tseng, W.-L.; Chen, H.-H.; Tseng, Y.-H.: Mining browsing behaviors for objectionable content filtering (2015) 0.00
    3.3322079E-4 = product of:
      0.00866374 = sum of:
        0.00866374 = weight(_text_:5 in 1818) [ClassicSimilarity], result of:
          0.00866374 = score(doc=1818,freq=2.0), product of:
            0.05374404 = queryWeight, product of:
              2.9180994 = idf(docFreq=6494, maxDocs=44218)
              0.018417481 = queryNorm
            0.16120374 = fieldWeight in 1818, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.9180994 = idf(docFreq=6494, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1818)
      0.03846154 = coord(1/26)
    
    Source
    Journal of the Association for Information Science and Technology. 66(2015) no.5, S.930-942
  5. Bian, G.-W.; Chen, H.-H.: Cross-language information access to multilingual collections on the Internet (2000) 0.00
    1.9194727E-4 = product of:
      0.004990629 = sum of:
        0.004990629 = product of:
          0.014971886 = sum of:
            0.014971886 = weight(_text_:22 in 4436) [ClassicSimilarity], result of:
              0.014971886 = score(doc=4436,freq=2.0), product of:
                0.06449488 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.018417481 = queryNorm
                0.23214069 = fieldWeight in 4436, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4436)
          0.33333334 = coord(1/3)
      0.03846154 = coord(1/26)
    
    Date
    16. 2.2000 14:22:39
  6. Ku, L.-W.; Ho, H.-W.; Chen, H.-H.: Opinion mining and relationship discovery using CopeOpi opinion analysis system (2009) 0.00
    1.5995605E-4 = product of:
      0.0041588573 = sum of:
        0.0041588573 = product of:
          0.012476572 = sum of:
            0.012476572 = weight(_text_:22 in 2938) [ClassicSimilarity], result of:
              0.012476572 = score(doc=2938,freq=2.0), product of:
                0.06449488 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.018417481 = queryNorm
                0.19345059 = fieldWeight in 2938, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2938)
          0.33333334 = coord(1/3)
      0.03846154 = coord(1/26)
    
    Abstract
    We present CopeOpi, an opinion-analysis system, which extracts from the Web opinions about specific targets, summarizes the polarity and strength of these opinions, and tracks opinion variations over time. Objects that yield similar opinion tendencies over a certain time period may be correlated due to the latent causal events. CopeOpi discovers relationships among objects based on their opinion-tracking plots and collocations. Event bursts are detected from the tracking plots, and the strength of opinion relationships is determined by the coverage of these plots. To evaluate opinion mining, we use the NTCIR corpus annotated with opinion information at sentence and document levels. CopeOpi achieves sentence- and document-level f-measures of 62% and 74%. For relationship discovery, we collected 1.3M economics-related documents from 93 Web sources over 22 months, and analyzed collocation-based, opinion-based, and hybrid models. We consider as correlated company pairs that demonstrate similar stock-price variations, and selected these as the gold standard for evaluation. Results show that opinion-based and collocation-based models complement each other, and that integrated models perform the best. The top 25, 50, and 100 pairs discovered achieve precision rates of 1, 0.92, and 0.79, respectively.