Search (12 results, page 1 of 1)

  • × author_ss:"Yang, C.C."
  1. Yang, C.C.; Luk, J.: Automatic generation of English/Chinese thesaurus based on a parallel corpus in laws (2003) 0.05
    0.05320441 = product of:
      0.10640882 = sum of:
        0.046679016 = weight(_text_:computer in 1616) [ClassicSimilarity], result of:
          0.046679016 = score(doc=1616,freq=8.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.28263903 = fieldWeight in 1616, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1616)
        0.04901477 = weight(_text_:network in 1616) [ClassicSimilarity], result of:
          0.04901477 = score(doc=1616,freq=4.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.24354391 = fieldWeight in 1616, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1616)
        0.010715035 = product of:
          0.02143007 = sum of:
            0.02143007 = weight(_text_:22 in 1616) [ClassicSimilarity], result of:
              0.02143007 = score(doc=1616,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.1354154 = fieldWeight in 1616, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1616)
          0.5 = coord(1/2)
      0.5 = coord(3/6)
    
    Abstract
    The information available in languages other than English in the World Wide Web is increasing significantly. According to a report from Computer Economics in 1999, 54% of Internet users are English speakers ("English Will Dominate Web for Only Three More Years," Computer Economics, July 9, 1999, http://www.computereconomics. com/new4/pr/pr990610.html). However, it is predicted that there will be only 60% increase in Internet users among English speakers verses a 150% growth among nonEnglish speakers for the next five years. By 2005, 57% of Internet users will be non-English speakers. A report by CNN.com in 2000 showed that the number of Internet users in China had been increased from 8.9 million to 16.9 million from January to June in 2000 ("Report: China Internet users double to 17 million," CNN.com, July, 2000, http://cnn.org/2000/TECH/computing/07/27/ china.internet.reut/index.html). According to Nielsen/ NetRatings, there was a dramatic leap from 22.5 millions to 56.6 millions Internet users from 2001 to 2002. China had become the second largest global at-home Internet population in 2002 (US's Internet population was 166 millions) (Robyn Greenspan, "China Pulls Ahead of Japan," Internet.com, April 22, 2002, http://cyberatias.internet.com/big-picture/geographics/article/0,,5911_1013841,00. html). All of the evidences reveal the importance of crosslingual research to satisfy the needs in the near future. Digital library research has been focusing in structural and semantic interoperability in the past. Searching and retrieving objects across variations in protocols, formats and disciplines are widely explored (Schatz, B., & Chen, H. (1999). Digital libraries: technological advances and social impacts. IEEE Computer, Special Issue an Digital Libraries, February, 32(2), 45-50.; Chen, H., Yen, J., & Yang, C.C. (1999). International activities: development of Asian digital libraries. IEEE Computer, Special Issue an Digital Libraries, 32(2), 48-49.). However, research in crossing language boundaries, especially across European languages and Oriental languages, is still in the initial stage. In this proposal, we put our focus an cross-lingual semantic interoperability by developing automatic generation of a cross-lingual thesaurus based an English/Chinese parallel corpus. When the searchers encounter retrieval problems, Professional librarians usually consult the thesaurus to identify other relevant vocabularies. In the problem of searching across language boundaries, a cross-lingual thesaurus, which is generated by co-occurrence analysis and Hopfield network, can be used to generate additional semantically relevant terms that cannot be obtained from dictionary. In particular, the automatically generated cross-lingual thesaurus is able to capture the unknown words that do not exist in a dictionary, such as names of persons, organizations, and events. Due to Hong Kong's unique history background, both English and Chinese are used as official languages in all legal documents. Therefore, English/Chinese cross-lingual information retrieval is critical for applications in courts and the government. In this paper, we develop an automatic thesaurus by the Hopfield network based an a parallel corpus collected from the Web site of the Department of Justice of the Hong Kong Special Administrative Region (HKSAR) Government. Experiments are conducted to measure the precision and recall of the automatic generated English/Chinese thesaurus. The result Shows that such thesaurus is a promising tool to retrieve relevant terms, especially in the language that is not the same as the input term. The direct translation of the input term can also be retrieved in most of the cases.
  2. Tang, X.; Yang, C.C.; Song, M.: Understanding the evolution of multiple scientific research domains using a content and network approach (2013) 0.01
    0.014292996 = product of:
      0.08575798 = sum of:
        0.08575798 = weight(_text_:network in 744) [ClassicSimilarity], result of:
          0.08575798 = score(doc=744,freq=6.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.42611307 = fieldWeight in 744, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.0390625 = fieldNorm(doc=744)
      0.16666667 = coord(1/6)
    
    Abstract
    Interdisciplinary research has been attracting more attention in recent decades. In this article, we compare the similarity between scientific research domains and quantifying the temporal similarities of domains. We narrowed our study to three research domains: information retrieval (IR), database (DB), and World Wide Web (W3), because the rapid development of the W3 domain substantially attracted research efforts from both IR and DB domains and introduced new research questions to these two areas. Most existing approaches either employed a content-based technique or a cocitation or coauthorship network-based technique to study the development trend of a research area. In this work, we proposed an effective way to quantify the similarities among different research domains by incorporating content similarity and coauthorship network similarity. Experimental results on DBLP (DataBase systems and Logic Programming) data related to IR, DB, and W3 domains showed that the W3 domain was getting closer to both IR and DB whereas the distance between IR and DB remained relatively constant. In addition, comparing to IR and W3 with the DB domain, the DB domain was more conservative and evolved relatively slower.
  3. Zhang, M.; Yang, C.C.: Using content and network analysis to understand the social support exchange patterns and user behaviors of an online smoking cessation intervention program (2015) 0.01
    0.011670183 = product of:
      0.0700211 = sum of:
        0.0700211 = weight(_text_:network in 1668) [ClassicSimilarity], result of:
          0.0700211 = score(doc=1668,freq=4.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.34791988 = fieldWeight in 1668, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1668)
      0.16666667 = coord(1/6)
    
    Abstract
    Informational support and nurturant support are two basic types of social support offered in online health communities. This study identifies types of social support in the QuitStop forum and brings insights to exchange patterns of social support and user behaviors with content analysis and social network analysis. Motivated by user information behavior, this study defines two patterns to describe social support exchange: initiated support exchange and invited support exchange. It is found that users with a longer quitting time tend to actively give initiated support, and recent quitters with a shorter abstinent time are likely to seek and receive invited support. This study also finds that support givers of informational support quit longer ago than support givers of nurturant support, and support receivers of informational support quit more recently than support receivers of nurturant support. Usually, informational support is offered by users at late quit stages to users at early quit stages. Nurturant support is also exchanged among users within the same quit stage. These findings help us understand how health consumers are supporting each other and reveal new capabilities of online intervention programs that can be designed to offer social support in a timely and effective manner.
  4. Chuang, K.Y.; Yang, C.C.: Informational support exchanges using different computer-mediated communication formats in a social media alcoholism community (2014) 0.01
    0.007858822 = product of:
      0.04715293 = sum of:
        0.04715293 = weight(_text_:computer in 1179) [ClassicSimilarity], result of:
          0.04715293 = score(doc=1179,freq=4.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.28550854 = fieldWeight in 1179, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1179)
      0.16666667 = coord(1/6)
    
    Abstract
    E-patients seeking information online often seek specific advice related to coping with their health condition(s) among social networking sites. They may be looking for social connectivity with compassionate strangers who may have experienced similar situations to share opinions and experiences rather than for authoritative medical information. Previous studies document distinct technological features and different levels of social support interaction patterns. It is expected that the design of the social media functions will have an impact on the user behavior of social support exchange. In this part of a multipart study, we investigate the social support types, in particular information support types, across multiple computer-mediated communication formats (forum, journal, and notes) within an alcoholism community using descriptive content analysis on 3 months of data from a MedHelp online peer support community. We present the results of identified informational support types including advice, referral, fact, personal experiences, and opinions, either offered or requested. Fact type was exchanged most often among the messages; however, there were some different patterns between notes and journal posts. Notes were used for maintaining relationships rather than as a main source for seeking information. Notes were similar to comments made to journal posts, which may indicate the friendship between journal readers and the author. These findings suggest that users may have initially joined the MedHelp Alcoholism Community for information-seeking purposes but continue participation even after they have completed with information gathering because of the relationships they formed with community members through social media features.
  5. Li, K.W.; Yang, C.C.: Automatic crosslingual thesaurus generated from the Hong Kong SAR Police Department Web Corpus for Crime Analysis (2005) 0.01
    0.006601652 = product of:
      0.039609913 = sum of:
        0.039609913 = weight(_text_:network in 3391) [ClassicSimilarity], result of:
          0.039609913 = score(doc=3391,freq=2.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.1968132 = fieldWeight in 3391, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.03125 = fieldNorm(doc=3391)
      0.16666667 = coord(1/6)
    
    Abstract
    For the sake of national security, very large volumes of data and information are generated and gathered daily. Much of this data and information is written in different languages, stored in different locations, and may be seemingly unconnected. Crosslingual semantic interoperability is a major challenge to generate an overview of this disparate data and information so that it can be analyzed, shared, searched, and summarized. The recent terrorist attacks and the tragic events of September 11, 2001 have prompted increased attention an national security and criminal analysis. Many Asian countries and cities, such as Japan, Taiwan, and Singapore, have been advised that they may become the next targets of terrorist attacks. Semantic interoperability has been a focus in digital library research. Traditional information retrieval (IR) approaches normally require a document to share some common keywords with the query. Generating the associations for the related terms between the two term spaces of users and documents is an important issue. The problem can be viewed as the creation of a thesaurus. Apart from this, terrorists and criminals may communicate through letters, e-mails, and faxes in languages other than English. The translation ambiguity significantly exacerbates the retrieval problem. The problem is expanded to crosslingual semantic interoperability. In this paper, we focus an the English/Chinese crosslingual semantic interoperability problem. However, the developed techniques are not limited to English and Chinese languages but can be applied to many other languages. English and Chinese are popular languages in the Asian region. Much information about national security or crime is communicated in these languages. An efficient automatically generated thesaurus between these languages is important to crosslingual information retrieval between English and Chinese languages. To facilitate crosslingual information retrieval, a corpus-based approach uses the term co-occurrence statistics in parallel or comparable corpora to construct a statistical translation model to cross the language boundary. In this paper, the text based approach to align English/Chinese Hong Kong Police press release documents from the Web is first presented. We also introduce an algorithmic approach to generate a robust knowledge base based an statistical correlation analysis of the semantics (knowledge) embedded in the bilingual press release corpus. The research output consisted of a thesaurus-like, semantic network knowledge base, which can aid in semanticsbased crosslingual information management and retrieval.
  6. Yang, C.C.; Chung, A.: ¬A personal agent for Chinese financial news on the Web (2002) 0.01
    0.0056084236 = product of:
      0.03365054 = sum of:
        0.03365054 = weight(_text_:services in 205) [ClassicSimilarity], result of:
          0.03365054 = score(doc=205,freq=2.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.2028165 = fieldWeight in 205, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.0390625 = fieldNorm(doc=205)
      0.16666667 = coord(1/6)
    
    Abstract
    As the Web has become a major channel of information dissemination, many newspapers expand their services by providing electronic versions of news information on the Web. However, most investors find it difficult to search for the financial information of interest from the huge Web information space-information overloading problem. In this article, we present a personal agent that utilizes user profiles and user relevance feedback to search for the Chinese Web financial news articles on behalf of users. A Chinese indexing component is developed to index the continuously fetched Chinese financial news articles. User profiles capture the basic knowledge of user preferences based on the sources of news articles, the regions of the news reported, categories of industries related, the listed companies, and user-specified keywords. User feedback captures the semantics of the user rated news articles. The search engine ranks the top 20 news articles that users are most interested in and report to the user daily or on demand. Experiments are conducted to measure the performance of the agents based on the inputs from user profiles and user feedback. It shows that simply using the user profiles does not increase the precision of the retrieval. However, user relevance feedback helps to increase the performance of the retrieval as the user interact with the system until it reaches the optimal performance. Combining both user profiles and user relevance feedback produces the best performance
  7. Lam, W.; Yang, C.C.; Menczer, F.: Introduction to the special topic section on mining Web resources for enhancing information retrieval (2007) 0.01
    0.005488624 = product of:
      0.03293174 = sum of:
        0.03293174 = product of:
          0.06586348 = sum of:
            0.06586348 = weight(_text_:resources in 600) [ClassicSimilarity], result of:
              0.06586348 = score(doc=600,freq=4.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.39925572 = fieldWeight in 600, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=600)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Footnote
    Einführung in einen Themenschwerpunkt "Mining Web resources for enhancing information retrieval"
  8. Wang, F.L.; Yang, C.C.: Mining Web data for Chinese segmentation (2007) 0.00
    0.0039204457 = product of:
      0.023522673 = sum of:
        0.023522673 = product of:
          0.047045346 = sum of:
            0.047045346 = weight(_text_:resources in 604) [ClassicSimilarity], result of:
              0.047045346 = score(doc=604,freq=4.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.28518265 = fieldWeight in 604, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=604)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    Modern information retrieval systems use keywords within documents as indexing terms for search of relevant documents. As Chinese is an ideographic character-based language, the words in the texts are not delimited by white spaces. Indexing of Chinese documents is impossible without a proper segmentation algorithm. Many Chinese segmentation algorithms have been proposed in the past. Traditional segmentation algorithms cannot operate without a large dictionary or a large corpus of training data. Nowadays, the Web has become the largest corpus that is ideal for Chinese segmentation. Although most search engines have problems in segmenting texts into proper words, they maintain huge databases of documents and frequencies of character sequences in the documents. Their databases are important potential resources for segmentation. In this paper, we propose a segmentation algorithm by mining Web data with the help of search engines. On the other hand, the Romanized pinyin of Chinese language indicates boundaries of words in the text. Our algorithm is the first to utilize the Romanized pinyin to segmentation. It is the first unified segmentation algorithm for the Chinese language from different geographical areas, and it is also domain independent because of the nature of the Web. Experiments have been conducted on the datasets of a recent Chinese segmentation competition. The results show that our algorithm outperforms the traditional algorithms in terms of precision and recall. Moreover, our algorithm can effectively deal with the problems of segmentation ambiguity, new word (unknown word) detection, and stop words.
    Footnote
    Beitrag eines Themenschwerpunktes "Mining Web resources for enhancing information retrieval"
  9. Shi, X.; Yang, C.C.: Mining related queries from Web search engine query logs using an improved association rule mining model (2007) 0.00
    0.0027721736 = product of:
      0.016633041 = sum of:
        0.016633041 = product of:
          0.033266082 = sum of:
            0.033266082 = weight(_text_:resources in 597) [ClassicSimilarity], result of:
              0.033266082 = score(doc=597,freq=2.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.20165458 = fieldWeight in 597, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=597)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Footnote
    Beitrag eines Themenschwerpunktes "Mining Web resources for enhancing information retrieval"
  10. Yang, C.C.; Lin, J.; Wei, C.-P.: Retaining knowledge for document management : category-tree integration by exploiting category relationships and hierarchical structures (2010) 0.00
    0.0027721736 = product of:
      0.016633041 = sum of:
        0.016633041 = product of:
          0.033266082 = sum of:
            0.033266082 = weight(_text_:resources in 3581) [ClassicSimilarity], result of:
              0.033266082 = score(doc=3581,freq=2.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.20165458 = fieldWeight in 3581, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3581)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    The category-tree document-classification structure is widely used by enterprises and information providers to organize, archive, and access documents for effective knowledge management. However, category trees from various sources use different hierarchical structures, which usually make mappings between categories in different category trees difficult. In this work, we propose a category-tree integration technique. We develop a method to learn the relationships between any two categories and develop operations such as mapping, splitting, and insertion for this integration. According to the parent-child relationship of the integrating categories, the developed decision rules use integration operations to integrate categories from the source category tree with those from the master category tree. A unified category tree can accumulate knowledge from multiple resources without forfeiting the knowledge in individual category trees. Experiments have been conducted to measure the performance of the integration operations and the accuracy of the integrated category trees. The proposed category-tree integration technique achieves greater than 80% integration accuracy, and the insert operation is the most frequently utilized, followed by map and split. The insert operation achieves 77% of F1 while the map and split operations achieves 86% and 29% of F1, respectively.
  11. Chau, M.; Lu, Y.; Fang, X.; Yang, C.C.: Characteristics of character usage in Chinese Web searching (2009) 0.00
    0.002551199 = product of:
      0.015307193 = sum of:
        0.015307193 = product of:
          0.030614385 = sum of:
            0.030614385 = weight(_text_:22 in 2456) [ClassicSimilarity], result of:
              0.030614385 = score(doc=2456,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.19345059 = fieldWeight in 2456, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2456)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Date
    22.11.2008 17:57:22
  12. Yang, C.C.; Liu, N.: Web site topic-hierarchy generation based on link structure (2009) 0.00
    0.002551199 = product of:
      0.015307193 = sum of:
        0.015307193 = product of:
          0.030614385 = sum of:
            0.030614385 = weight(_text_:22 in 2738) [ClassicSimilarity], result of:
              0.030614385 = score(doc=2738,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.19345059 = fieldWeight in 2738, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2738)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Date
    22. 3.2009 12:51:47