Search (61 results, page 1 of 4)

  • × author_ss:"Chen, H."
  1. Huang, C.; Fu, T.; Chen, H.: Text-based video content classification for online video-sharing sites (2010) 0.05
    0.04860077 = product of:
      0.19440308 = sum of:
        0.04887987 = weight(_text_:web in 3452) [ClassicSimilarity], result of:
          0.04887987 = score(doc=3452,freq=14.0), product of:
            0.10247572 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.031400457 = queryNorm
            0.47698978 = fieldWeight in 3452, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3452)
        0.116696276 = weight(_text_:2.0 in 3452) [ClassicSimilarity], result of:
          0.116696276 = score(doc=3452,freq=8.0), product of:
            0.18211427 = queryWeight, product of:
              5.799733 = idf(docFreq=363, maxDocs=44218)
              0.031400457 = queryNorm
            0.640786 = fieldWeight in 3452, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              5.799733 = idf(docFreq=363, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3452)
        0.01956797 = product of:
          0.03913594 = sum of:
            0.03913594 = weight(_text_:online in 3452) [ClassicSimilarity], result of:
              0.03913594 = score(doc=3452,freq=12.0), product of:
                0.09529729 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.031400457 = queryNorm
                0.41067213 = fieldWeight in 3452, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3452)
          0.5 = coord(1/2)
        0.009258964 = weight(_text_:information in 3452) [ClassicSimilarity], result of:
          0.009258964 = score(doc=3452,freq=6.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.16796975 = fieldWeight in 3452, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3452)
      0.25 = coord(4/16)
    
    Abstract
    With the emergence of Web 2.0, sharing personal content, communicating ideas, and interacting with other online users in Web 2.0 communities have become daily routines for online users. User-generated data from Web 2.0 sites provide rich personal information (e.g., personal preferences and interests) and can be utilized to obtain insight about cyber communities and their social networks. Many studies have focused on leveraging user-generated information to analyze blogs and forums, but few studies have applied this approach to video-sharing Web sites. In this study, we propose a text-based framework for video content classification of online-video sharing Web sites. Different types of user-generated data (e.g., titles, descriptions, and comments) were used as proxies for online videos, and three types of text features (lexical, syntactic, and content-specific features) were extracted. Three feature-based classification techniques (C4.5, Naïve Bayes, and Support Vector Machine) were used to classify videos. To evaluate the proposed framework, user-generated data from candidate videos, which were identified by searching user-given keywords on YouTube, were first collected. Then, a subset of the collected data was randomly selected and manually tagged by users as our experiment data. The experimental results showed that the proposed approach was able to classify online videos based on users' interests with accuracy rates up to 87.2%, and all three types of text features contributed to discriminating videos. Support Vector Machine outperformed C4.5 and Naïve Bayes techniques in our experiments. In addition, our case study further demonstrated that accurate video-classification results are very useful for identifying implicit cyber communities on video-sharing Web sites.
    Object
    Web 2.0
    Source
    Journal of the American Society for Information Science and Technology. 61(2010) no.5, S.891-906
  2. Chen, H.: Introduction to the JASIST special topic section on Web retrieval and mining : A machine learning perspective (2003) 0.03
    0.03444291 = product of:
      0.13777164 = sum of:
        0.040864702 = weight(_text_:wide in 1610) [ClassicSimilarity], result of:
          0.040864702 = score(doc=1610,freq=2.0), product of:
            0.13912784 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.031400457 = queryNorm
            0.29372054 = fieldWeight in 1610, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046875 = fieldNorm(doc=1610)
        0.04957324 = weight(_text_:web in 1610) [ClassicSimilarity], result of:
          0.04957324 = score(doc=1610,freq=10.0), product of:
            0.10247572 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.031400457 = queryNorm
            0.48375595 = fieldWeight in 1610, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=1610)
        0.014343925 = weight(_text_:information in 1610) [ClassicSimilarity], result of:
          0.014343925 = score(doc=1610,freq=10.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.2602176 = fieldWeight in 1610, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=1610)
        0.032989766 = weight(_text_:retrieval in 1610) [ClassicSimilarity], result of:
          0.032989766 = score(doc=1610,freq=6.0), product of:
            0.09498371 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.031400457 = queryNorm
            0.34732026 = fieldWeight in 1610, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=1610)
      0.25 = coord(4/16)
    
    Abstract
    Research in information retrieval (IR) has advanced significantly in the past few decades. Many tasks, such as indexing and text categorization, can be performed automatically with minimal human effort. Machine learning has played an important role in such automation by learning various patterns such as document topics, text structures, and user interests from examples. In recent years, it has become increasingly difficult to search for useful information an the World Wide Web because of its large size and unstructured nature. Useful information and resources are often hidden in the Web. While machine learning has been successfully applied to traditional IR systems, it poses some new challenges to apply these algorithms to the Web due to its large size, link structure, diversity in content and languages, and dynamic nature. On the other hand, such characteristics of the Web also provide interesting patterns and knowledge that do not present in traditional information retrieval systems.
    Source
    Journal of the American Society for Information Science and technology. 54(2003) no.7, S.621-624
  3. Dumais, S.; Chen, H.: Hierarchical classification of Web content (2000) 0.03
    0.03255348 = product of:
      0.13021392 = sum of:
        0.044339646 = weight(_text_:web in 492) [ClassicSimilarity], result of:
          0.044339646 = score(doc=492,freq=2.0), product of:
            0.10247572 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.031400457 = queryNorm
            0.43268442 = fieldWeight in 492, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.09375 = fieldNorm(doc=492)
        0.019172618 = product of:
          0.038345236 = sum of:
            0.038345236 = weight(_text_:online in 492) [ClassicSimilarity], result of:
              0.038345236 = score(doc=492,freq=2.0), product of:
                0.09529729 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.031400457 = queryNorm
                0.40237486 = fieldWeight in 492, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.09375 = fieldNorm(doc=492)
          0.5 = coord(1/2)
        0.012829596 = weight(_text_:information in 492) [ClassicSimilarity], result of:
          0.012829596 = score(doc=492,freq=2.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.23274569 = fieldWeight in 492, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=492)
        0.05387206 = weight(_text_:retrieval in 492) [ClassicSimilarity], result of:
          0.05387206 = score(doc=492,freq=4.0), product of:
            0.09498371 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.031400457 = queryNorm
            0.5671716 = fieldWeight in 492, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.09375 = fieldNorm(doc=492)
      0.25 = coord(4/16)
    
    Source
    Proceedings of ACM SIGIR 23rd International Conference on Research and Development in Information Retrieval. Ed. by N.J. Belkin, P. Ingwersen u. M.K. Leong
    Theme
    Klassifikationssysteme im Online-Retrieval
  4. Chau, M.; Shiu, B.; Chan, M.; Chen, H.: Redips: backlink search and analysis on the Web for business intelligence analysis (2007) 0.02
    0.024138859 = product of:
      0.096555434 = sum of:
        0.03405392 = weight(_text_:wide in 142) [ClassicSimilarity], result of:
          0.03405392 = score(doc=142,freq=2.0), product of:
            0.13912784 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.031400457 = queryNorm
            0.24476713 = fieldWeight in 142, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.0390625 = fieldNorm(doc=142)
        0.045253962 = weight(_text_:web in 142) [ClassicSimilarity], result of:
          0.045253962 = score(doc=142,freq=12.0), product of:
            0.10247572 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.031400457 = queryNorm
            0.4416067 = fieldWeight in 142, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=142)
        0.00798859 = product of:
          0.01597718 = sum of:
            0.01597718 = weight(_text_:online in 142) [ClassicSimilarity], result of:
              0.01597718 = score(doc=142,freq=2.0), product of:
                0.09529729 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.031400457 = queryNorm
                0.16765618 = fieldWeight in 142, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=142)
          0.5 = coord(1/2)
        0.009258964 = weight(_text_:information in 142) [ClassicSimilarity], result of:
          0.009258964 = score(doc=142,freq=6.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.16796975 = fieldWeight in 142, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=142)
      0.25 = coord(4/16)
    
    Abstract
    The World Wide Web presents significant opportunities for business intelligence analysis as it can provide information about a company's external environment and its stakeholders. Traditional business intelligence analysis on the Web has focused on simple keyword searching. Recently, it has been suggested that the incoming links, or backlinks, of a company's Web site (i.e., other Web pages that have a hyperlink pointing to the company of Interest) can provide important insights about the company's "online communities." Although analysis of these communities can provide useful signals for a company and information about its stakeholder groups, the manual analysis process can be very time-consuming for business analysts and consultants. In this article, we present a tool called Redips that automatically integrates backlink meta-searching and text-mining techniques to facilitate users in performing such business intelligence analysis on the Web. The architectural design and implementation of the tool are presented in the article. To evaluate the effectiveness, efficiency, and user satisfaction of Redips, an experiment was conducted to compare the tool with two popular business Intelligence analysis methods-using backlink search engines and manual browsing. The experiment results showed that Redips was statistically more effective than both benchmark methods (in terms of Recall and F-measure) but required more time in search tasks. In terms of user satisfaction, Redips scored statistically higher than backlink search engines in all five measures used, and also statistically higher than manual browsing in three measures.
    Source
    Journal of the American Society for Information Science and Technology. 58(2007) no.3, S.351-365
  5. Chen, H.; Chau, M.: Web mining : machine learning for Web applications (2003) 0.02
    0.023496704 = product of:
      0.12531576 = sum of:
        0.040864702 = weight(_text_:wide in 4242) [ClassicSimilarity], result of:
          0.040864702 = score(doc=4242,freq=2.0), product of:
            0.13912784 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.031400457 = queryNorm
            0.29372054 = fieldWeight in 4242, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046875 = fieldNorm(doc=4242)
        0.07010714 = weight(_text_:web in 4242) [ClassicSimilarity], result of:
          0.07010714 = score(doc=4242,freq=20.0), product of:
            0.10247572 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.031400457 = queryNorm
            0.6841342 = fieldWeight in 4242, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4242)
        0.014343925 = weight(_text_:information in 4242) [ClassicSimilarity], result of:
          0.014343925 = score(doc=4242,freq=10.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.2602176 = fieldWeight in 4242, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=4242)
      0.1875 = coord(3/16)
    
    Abstract
    With more than two billion pages created by millions of Web page authors and organizations, the World Wide Web is a tremendously rich knowledge base. The knowledge comes not only from the content of the pages themselves, but also from the unique characteristics of the Web, such as its hyperlink structure and its diversity of content and languages. Analysis of these characteristics often reveals interesting patterns and new knowledge. Such knowledge can be used to improve users' efficiency and effectiveness in searching for information an the Web, and also for applications unrelated to the Web, such as support for decision making or business management. The Web's size and its unstructured and dynamic content, as well as its multilingual nature, make the extraction of useful knowledge a challenging research problem. Furthermore, the Web generates a large amount of data in other formats that contain valuable information. For example, Web server logs' information about user access patterns can be used for information personalization or improving Web page design.
    Source
    Annual review of information science and technology. 38(2004), S.289-330
  6. Huang, Z.; Chung, Z.W.; Chen, H.: ¬A graph model for e-commerce recommender systems (2004) 0.02
    0.022802755 = product of:
      0.09121102 = sum of:
        0.022169823 = weight(_text_:web in 501) [ClassicSimilarity], result of:
          0.022169823 = score(doc=501,freq=2.0), product of:
            0.10247572 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.031400457 = queryNorm
            0.21634221 = fieldWeight in 501, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=501)
        0.016603975 = product of:
          0.03320795 = sum of:
            0.03320795 = weight(_text_:online in 501) [ClassicSimilarity], result of:
              0.03320795 = score(doc=501,freq=6.0), product of:
                0.09529729 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.031400457 = queryNorm
                0.34846687 = fieldWeight in 501, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.046875 = fieldNorm(doc=501)
          0.5 = coord(1/2)
        0.014343925 = weight(_text_:information in 501) [ClassicSimilarity], result of:
          0.014343925 = score(doc=501,freq=10.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.2602176 = fieldWeight in 501, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=501)
        0.0380933 = weight(_text_:retrieval in 501) [ClassicSimilarity], result of:
          0.0380933 = score(doc=501,freq=8.0), product of:
            0.09498371 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.031400457 = queryNorm
            0.40105087 = fieldWeight in 501, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=501)
      0.25 = coord(4/16)
    
    Abstract
    Information overload on the Web has created enormous challenges to customers selecting products for online purchases and to online businesses attempting to identify customers' preferences efficiently. Various recommender systems employing different data representations and recommendation methods are currently used to address these challenges. In this research, we developed a graph model that provides a generic data representation and can support different recommendation methods. To demonstrate its usefulness and flexibility, we developed three recommendation methods: direct retrieval, association mining, and high-degree association retrieval. We used a data set from an online bookstore as our research test-bed. Evaluation results showed that combining product content information and historical customer transaction information achieved more accurate predictions and relevant recommendations than using only collaborative information. However, comparisons among different methods showed that high-degree association retrieval did not perform significantly better than the association mining method or the direct retrieval method in our test-bed.
    Source
    Journal of the American Society for Information Science and technology. 55(2004) no.3, S.259-274
  7. Fu, T.; Abbasi, A.; Chen, H.: ¬A focused crawler for Dark Web forums (2010) 0.02
    0.021157756 = product of:
      0.084631026 = sum of:
        0.05542456 = weight(_text_:web in 3471) [ClassicSimilarity], result of:
          0.05542456 = score(doc=3471,freq=18.0), product of:
            0.10247572 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.031400457 = queryNorm
            0.5408555 = fieldWeight in 3471, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3471)
        0.00798859 = product of:
          0.01597718 = sum of:
            0.01597718 = weight(_text_:online in 3471) [ClassicSimilarity], result of:
              0.01597718 = score(doc=3471,freq=2.0), product of:
                0.09529729 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.031400457 = queryNorm
                0.16765618 = fieldWeight in 3471, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3471)
          0.5 = coord(1/2)
        0.005345665 = weight(_text_:information in 3471) [ClassicSimilarity], result of:
          0.005345665 = score(doc=3471,freq=2.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.09697737 = fieldWeight in 3471, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3471)
        0.015872208 = weight(_text_:retrieval in 3471) [ClassicSimilarity], result of:
          0.015872208 = score(doc=3471,freq=2.0), product of:
            0.09498371 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.031400457 = queryNorm
            0.16710453 = fieldWeight in 3471, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3471)
      0.25 = coord(4/16)
    
    Abstract
    The unprecedented growth of the Internet has given rise to the Dark Web, the problematic facet of the Web associated with cybercrime, hate, and extremism. Despite the need for tools to collect and analyze Dark Web forums, the covert nature of this part of the Internet makes traditional Web crawling techniques insufficient for capturing such content. In this study, we propose a novel crawling system designed to collect Dark Web forum content. The system uses a human-assisted accessibility approach to gain access to Dark Web forums. Several URL ordering features and techniques enable efficient extraction of forum postings. The system also includes an incremental crawler coupled with a recall-improvement mechanism intended to facilitate enhanced retrieval and updating of collected content. Experiments conducted to evaluate the effectiveness of the human-assisted accessibility approach and the recall-improvement-based, incremental-update procedure yielded favorable results. The human-assisted approach significantly improved access to Dark Web forums while the incremental crawler with recall improvement also outperformed standard periodic- and incremental-update approaches. Using the system, we were able to collect over 100 Dark Web forums from three regions. A case study encompassing link and content analysis of collected forums was used to illustrate the value and importance of gathering and analyzing content from such online communities.
    Source
    Journal of the American Society for Information Science and Technology. 61(2010) no.6, S.1213-1231
  8. Chen, H.; Lally, A.M.; Zhu, B.; Chau, M.: HelpfulMed : Intelligent searching for medical information over the Internet (2003) 0.02
    0.020422183 = product of:
      0.08168873 = sum of:
        0.03694971 = weight(_text_:web in 1615) [ClassicSimilarity], result of:
          0.03694971 = score(doc=1615,freq=8.0), product of:
            0.10247572 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.031400457 = queryNorm
            0.36057037 = fieldWeight in 1615, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1615)
        0.00798859 = product of:
          0.01597718 = sum of:
            0.01597718 = weight(_text_:online in 1615) [ClassicSimilarity], result of:
              0.01597718 = score(doc=1615,freq=2.0), product of:
                0.09529729 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.031400457 = queryNorm
                0.16765618 = fieldWeight in 1615, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1615)
          0.5 = coord(1/2)
        0.009258964 = weight(_text_:information in 1615) [ClassicSimilarity], result of:
          0.009258964 = score(doc=1615,freq=6.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.16796975 = fieldWeight in 1615, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1615)
        0.027491473 = weight(_text_:retrieval in 1615) [ClassicSimilarity], result of:
          0.027491473 = score(doc=1615,freq=6.0), product of:
            0.09498371 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.031400457 = queryNorm
            0.28943354 = fieldWeight in 1615, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1615)
      0.25 = coord(4/16)
    
    Abstract
    The Medical professionals and researchers need information from reputable sources to accomplish their work. Unfortunately, the Web has a large number of documents that are irrelevant to their work, even those documents that purport to be "medically-related." This paper describes an architecture designed to integrate advanced searching and indexing algorithms, an automatic thesaurus, or "concept space," and Kohonen-based Self-Organizing Map (SOM) technologies to provide searchers with finegrained results. Initial results indicate that these systems provide complementary retrieval functionalities. HelpfulMed not only allows users to search Web pages and other online databases, but also allows them to build searches through the use of an automatic thesaurus and browse a graphical display of medical-related topics. Evaluation results for each of the different components are included. Our spidering algorithm outperformed both breadth-first search and PageRank spiders an a test collection of 100,000 Web pages. The automatically generated thesaurus performed as well as both MeSH and UMLS-systems which require human mediation for currency. Lastly, a variant of the Kohonen SOM was comparable to MeSH terms in perceived cluster precision and significantly better at perceived cluster recall.
    Footnote
    Teil eines Themenheftes: "Web retrieval and mining: A machine learning perspective"
    Source
    Journal of the American Society for Information Science and technology. 54(2003) no.7, S.683-694
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
  9. Chung, W.; Chen, H.: Browsing the underdeveloped Web : an experiment on the Arabic Medical Web Directory (2009) 0.02
    0.019786317 = product of:
      0.10552702 = sum of:
        0.07993444 = weight(_text_:web in 2733) [ClassicSimilarity], result of:
          0.07993444 = score(doc=2733,freq=26.0), product of:
            0.10247572 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.031400457 = queryNorm
            0.780033 = fieldWeight in 2733, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=2733)
        0.012829596 = weight(_text_:information in 2733) [ClassicSimilarity], result of:
          0.012829596 = score(doc=2733,freq=8.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.23274569 = fieldWeight in 2733, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2733)
        0.012762985 = product of:
          0.02552597 = sum of:
            0.02552597 = weight(_text_:22 in 2733) [ClassicSimilarity], result of:
              0.02552597 = score(doc=2733,freq=2.0), product of:
                0.10995905 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.031400457 = queryNorm
                0.23214069 = fieldWeight in 2733, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2733)
          0.5 = coord(1/2)
      0.1875 = coord(3/16)
    
    Abstract
    While the Web has grown significantly in recent years, some portions of the Web remain largely underdeveloped, as shown in a lack of high-quality content and functionality. An example is the Arabic Web, in which a lack of well-structured Web directories limits users' ability to browse for Arabic resources. In this research, we proposed an approach to building Web directories for the underdeveloped Web and developed a proof-of-concept prototype called the Arabic Medical Web Directory (AMedDir) that supports browsing of over 5,000 Arabic medical Web sites and pages organized in a hierarchical structure. We conducted an experiment involving Arab participants and found that the AMedDir significantly outperformed two benchmark Arabic Web directories in terms of browsing effectiveness, efficiency, information quality, and user satisfaction. Participants expressed strong preference for the AMedDir and provided many positive comments. This research thus contributes to developing a useful Web directory for organizing the information in the Arabic medical domain and to a better understanding of how to support browsing on the underdeveloped Web.
    Date
    22. 3.2009 17:57:50
    Source
    Journal of the American Society for Information Science and Technology. 60(2009) no.3, S.595-607
    Theme
    Information Gateway
  10. Chen, H.; Dhar, V.: Cognitive process as a basis for intelligent retrieval system design (1991) 0.02
    0.016814362 = product of:
      0.0896766 = sum of:
        0.018076118 = product of:
          0.036152236 = sum of:
            0.036152236 = weight(_text_:online in 3845) [ClassicSimilarity], result of:
              0.036152236 = score(doc=3845,freq=4.0), product of:
                0.09529729 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.031400457 = queryNorm
                0.37936267 = fieldWeight in 3845, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3845)
          0.5 = coord(1/2)
        0.014814342 = weight(_text_:information in 3845) [ClassicSimilarity], result of:
          0.014814342 = score(doc=3845,freq=6.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.2687516 = fieldWeight in 3845, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=3845)
        0.05678614 = weight(_text_:retrieval in 3845) [ClassicSimilarity], result of:
          0.05678614 = score(doc=3845,freq=10.0), product of:
            0.09498371 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.031400457 = queryNorm
            0.59785134 = fieldWeight in 3845, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0625 = fieldNorm(doc=3845)
      0.1875 = coord(3/16)
    
    Abstract
    2 studies were conducted to investigate the cognitive processes involved in online document-based information retrieval. These studies led to the development of 5 computerised models of online document retrieval. These models were incorporated into a design of an 'intelligent' document-based retrieval system. Following a discussion of this system, discusses the broader implications of the research for the design of information retrieval sysems
    Source
    Information processing and management. 27(1991) no.5, S.405-432
  11. Qin, J.; Zhou, Y.; Chau, M.; Chen, H.: Multilingual Web retrieval : an experiment in English-Chinese business intelligence (2006) 0.02
    0.016678434 = product of:
      0.08895165 = sum of:
        0.045253962 = weight(_text_:web in 5054) [ClassicSimilarity], result of:
          0.045253962 = score(doc=5054,freq=12.0), product of:
            0.10247572 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.031400457 = queryNorm
            0.4416067 = fieldWeight in 5054, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5054)
        0.01195327 = weight(_text_:information in 5054) [ClassicSimilarity], result of:
          0.01195327 = score(doc=5054,freq=10.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.21684799 = fieldWeight in 5054, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5054)
        0.031744417 = weight(_text_:retrieval in 5054) [ClassicSimilarity], result of:
          0.031744417 = score(doc=5054,freq=8.0), product of:
            0.09498371 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.031400457 = queryNorm
            0.33420905 = fieldWeight in 5054, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5054)
      0.1875 = coord(3/16)
    
    Abstract
    As increasing numbers of non-English resources have become available on the Web, the interesting and important issue of how Web users can retrieve documents in different languages has arisen. Cross-language information retrieval (CLIP), the study of retrieving information in one language by queries expressed in another language, is a promising approach to the problem. Cross-language information retrieval has attracted much attention in recent years. Most research systems have achieved satisfactory performance on standard Text REtrieval Conference (TREC) collections such as news articles, but CLIR techniques have not been widely studied and evaluated for applications such as Web portals. In this article, the authors present their research in developing and evaluating a multilingual English-Chinese Web portal that incorporates various CLIP techniques for use in the business domain. A dictionary-based approach was adopted and combines phrasal translation, co-occurrence analysis, and pre- and posttranslation query expansion. The portal was evaluated by domain experts, using a set of queries in both English and Chinese. The experimental results showed that co-occurrence-based phrasal translation achieved a 74.6% improvement in precision over simple word-byword translation. When used together, pre- and posttranslation query expansion improved the performance slightly, achieving a 78.0% improvement over the baseline word-by-word translation approach. In general, applying CLIR techniques in Web applications shows promise.
    Footnote
    Beitrag einer special topic section on multilingual information systems
    Source
    Journal of the American Society for Information Science and Technology. 57(2006) no.5, S.671-683
  12. Zhu, B.; Chen, H.: Information visualization (2004) 0.01
    0.011846285 = product of:
      0.063180186 = sum of:
        0.012932398 = weight(_text_:web in 4276) [ClassicSimilarity], result of:
          0.012932398 = score(doc=4276,freq=2.0), product of:
            0.10247572 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.031400457 = queryNorm
            0.12619963 = fieldWeight in 4276, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4276)
        0.01714784 = weight(_text_:information in 4276) [ClassicSimilarity], result of:
          0.01714784 = score(doc=4276,freq=42.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.31108427 = fieldWeight in 4276, product of:
              6.4807405 = tf(freq=42.0), with freq of:
                42.0 = termFreq=42.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4276)
        0.03309995 = weight(_text_:software in 4276) [ClassicSimilarity], result of:
          0.03309995 = score(doc=4276,freq=6.0), product of:
            0.124570385 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.031400457 = queryNorm
            0.26571283 = fieldWeight in 4276, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4276)
      0.1875 = coord(3/16)
    
    Abstract
    Advanced technology has resulted in the generation of about one million terabytes of information every year. Ninety-reine percent of this is available in digital format (Keim, 2001). More information will be generated in the next three years than was created during all of previous human history (Keim, 2001). Collecting information is no longer a problem, but extracting value from information collections has become progressively more difficult. Various search engines have been developed to make it easier to locate information of interest, but these work well only for a person who has a specific goal and who understands what and how information is stored. This usually is not the Gase. Visualization was commonly thought of in terms of representing human mental processes (MacEachren, 1991; Miller, 1984). The concept is now associated with the amplification of these mental processes (Card, Mackinlay, & Shneiderman, 1999). Human eyes can process visual cues rapidly, whereas advanced information analysis techniques transform the computer into a powerful means of managing digitized information. Visualization offers a link between these two potent systems, the human eye and the computer (Gershon, Eick, & Card, 1998), helping to identify patterns and to extract insights from large amounts of information. The identification of patterns is important because it may lead to a scientific discovery, an interpretation of clues to solve a crime, the prediction of catastrophic weather, a successful financial investment, or a better understanding of human behavior in a computermediated environment. Visualization technology shows considerable promise for increasing the value of large-scale collections of information, as evidenced by several commercial applications of TreeMap (e.g., http://www.smartmoney.com) and Hyperbolic tree (e.g., http://www.inxight.com) to visualize large-scale hierarchical structures. Although the proliferation of visualization technologies dates from the 1990s where sophisticated hardware and software made increasingly faster generation of graphical objects possible, the role of visual aids in facilitating the construction of mental images has a long history. Visualization has been used to communicate ideas, to monitor trends implicit in data, and to explore large volumes of data for hypothesis generation. Imagine traveling to a strange place without a map, having to memorize physical and chemical properties of an element without Mendeleyev's periodic table, trying to understand the stock market without statistical diagrams, or browsing a collection of documents without interactive visual aids. A collection of information can lose its value simply because of the effort required for exhaustive exploration. Such frustrations can be overcome by visualization.
    Visualization can be classified as scientific visualization, software visualization, or information visualization. Although the data differ, the underlying techniques have much in common. They use the same elements (visual cues) and follow the same rules of combining visual cues to deliver patterns. They all involve understanding human perception (Encarnacao, Foley, Bryson, & Feiner, 1994) and require domain knowledge (Tufte, 1990). Because most decisions are based an unstructured information, such as text documents, Web pages, or e-mail messages, this chapter focuses an the visualization of unstructured textual documents. The chapter reviews information visualization techniques developed over the last decade and examines how they have been applied in different domains. The first section provides the background by describing visualization history and giving overviews of scientific, software, and information visualization as well as the perceptual aspects of visualization. The next section assesses important visualization techniques that convert abstract information into visual objects and facilitate navigation through displays an a computer screen. It also explores information analysis algorithms that can be applied to identify or extract salient visualizable structures from collections of information. Information visualization systems that integrate different types of technologies to address problems in different domains are then surveyed; and we move an to a survey and critique of visualization system evaluation studies. The chapter concludes with a summary and identification of future research directions.
    Source
    Annual review of information science and technology. 39(2005), S.139-177
  13. Fu, T.; Abbasi, A.; Chen, H.: ¬A hybrid approach to Web forum interactional coherence analysis (2008) 0.01
    0.010698275 = product of:
      0.057057466 = sum of:
        0.038399264 = weight(_text_:web in 1872) [ClassicSimilarity], result of:
          0.038399264 = score(doc=1872,freq=6.0), product of:
            0.10247572 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.031400457 = queryNorm
            0.37471575 = fieldWeight in 1872, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=1872)
        0.009586309 = product of:
          0.019172618 = sum of:
            0.019172618 = weight(_text_:online in 1872) [ClassicSimilarity], result of:
              0.019172618 = score(doc=1872,freq=2.0), product of:
                0.09529729 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.031400457 = queryNorm
                0.20118743 = fieldWeight in 1872, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1872)
          0.5 = coord(1/2)
        0.009071894 = weight(_text_:information in 1872) [ClassicSimilarity], result of:
          0.009071894 = score(doc=1872,freq=4.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.16457605 = fieldWeight in 1872, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=1872)
      0.1875 = coord(3/16)
    
    Abstract
    Despite the rapid growth of text-based computer-mediated communication (CMC), its limitations have rendered the media highly incoherent. This poses problems for content analysis of online discourse archives. Interactional coherence analysis (ICA) attempts to accurately identify and construct CMC interaction networks. In this study, we propose the Hybrid Interactional Coherence (HIC) algorithm for identification of web forum interaction. HIC utilizes a bevy of system and linguistic features, including message header information, quotations, direct address, and lexical relations. Furthermore, several similarity-based methods including a Lexical Match Algorithm (LMA) and a sliding window method are utilized to account for interactional idiosyncrasies. Experiments results on two web forums revealed that the proposed HIC algorithm significantly outperformed comparison techniques in terms of precision, recall, and F-measure at both the forum and thread levels. Additionally, an example was used to illustrate how the improved ICA results can facilitate enhanced social network and role analysis capabilities.
    Source
    Journal of the American Society for Information Science and Technology. 59(2008) no.8, S.1195-1209
  14. Chen, H.; Chung, W.; Qin, J.; Reid, E.; Sageman, M.; Weimann, G.: Uncovering the dark Web : a case study of Jihad on the Web (2008) 0.01
    0.010435179 = product of:
      0.08348143 = sum of:
        0.06650947 = weight(_text_:web in 1880) [ClassicSimilarity], result of:
          0.06650947 = score(doc=1880,freq=18.0), product of:
            0.10247572 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.031400457 = queryNorm
            0.64902663 = fieldWeight in 1880, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=1880)
        0.01697196 = weight(_text_:information in 1880) [ClassicSimilarity], result of:
          0.01697196 = score(doc=1880,freq=14.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.3078936 = fieldWeight in 1880, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=1880)
      0.125 = coord(2/16)
    
    Abstract
    While the Web has become a worldwide platform for communication, terrorists share their ideology and communicate with members on the Dark Web - the reverse side of the Web used by terrorists. Currently, the problems of information overload and difficulty to obtain a comprehensive picture of terrorist activities hinder effective and efficient analysis of terrorist information on the Web. To improve understanding of terrorist activities, we have developed a novel methodology for collecting and analyzing Dark Web information. The methodology incorporates information collection, analysis, and visualization techniques, and exploits various Web information sources. We applied it to collecting and analyzing information of 39 Jihad Web sites and developed visualization of their site contents, relationships, and activity levels. An expert evaluation showed that the methodology is very useful and promising, having a high potential to assist in investigation and understanding of terrorist activities by producing results that could potentially help guide both policymaking and intelligence research.
    Source
    Journal of the American Society for Information Science and Technology. 59(2008) no.8, S.1347-1359
  15. Chen, H.; Baptista Nunes, J.M.; Ragsdell, G.; An, X.: Somatic and cultural knowledge : drivers of a habitus-driven model of tacit knowledge acquisition (2019) 0.01
    0.01023857 = product of:
      0.054605708 = sum of:
        0.023837745 = weight(_text_:wide in 5460) [ClassicSimilarity], result of:
          0.023837745 = score(doc=5460,freq=2.0), product of:
            0.13912784 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.031400457 = queryNorm
            0.171337 = fieldWeight in 5460, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.02734375 = fieldNorm(doc=5460)
        0.0037419656 = weight(_text_:information in 5460) [ClassicSimilarity], result of:
          0.0037419656 = score(doc=5460,freq=2.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.06788416 = fieldWeight in 5460, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02734375 = fieldNorm(doc=5460)
        0.027025996 = weight(_text_:software in 5460) [ClassicSimilarity], result of:
          0.027025996 = score(doc=5460,freq=4.0), product of:
            0.124570385 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.031400457 = queryNorm
            0.21695362 = fieldWeight in 5460, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02734375 = fieldNorm(doc=5460)
      0.1875 = coord(3/16)
    
    Abstract
    The purpose of this paper is to identify and explain the role of individual learning and development in acquiring tacit knowledge in the context of the inexorable and intense continuous change (technological and otherwise) that characterizes our society today, and also to investigate the software (SW) sector, which is at the core of contemporary continuous change and is a paradigm of effective and intrinsic knowledge sharing (KS). This makes the SW sector unique and different from others where KS is so hard to implement. Design/methodology/approach The study employed an inductive qualitative approach based on a multi-case study approach, composed of three successful SW companies in China. These companies are representative of the fabric of the sector, namely a small- and medium-sized enterprise, a large private company and a large state-owned enterprise. The fieldwork included 44 participants who were interviewed using a semi-structured script. The interview data were coded and interpreted following the Straussian grounded theory pattern of open coding, axial coding and selective coding. The process of interviewing was stopped when theoretical saturation was achieved after a careful process of theoretical sampling.
    Findings The findings of this research suggest that individual learning and development are deemed to be the fundamental feature for professional success and survival in the continuously changing environment of the SW industry today. However, individual learning was described by the participants as much more than a mere individual process. It involves a collective and participatory effort within the organization and the sector as a whole, and a KS process that transcends organizational, cultural and national borders. Individuals in particular are mostly motivated by the pressing need to face and adapt to the dynamic and changeable environments of today's digital society that is led by the sector. Software practitioners are continuously in need of learning, refreshing and accumulating tacit knowledge, partly because it is required by their companies, but also due to a sound awareness of continuous technical and technological changes that seem only to increase with the advances of information technology. This led to a clear theoretical understanding that the continuous change that faces the sector has led to individual acquisition of culture and somatic knowledge that in turn lay the foundation for not only the awareness of the need for continuous individual professional development but also for the creation of habitus related to KS and continuous learning. Originality/value The study reported in this paper shows that there is a theoretical link between the existence of conducive organizational and sector-wide somatic and cultural knowledge, and the success of KS practices that lead to individual learning and development. Therefore, the theory proposed suggests that somatic and cultural knowledge are crucial drivers for the creation of habitus of individual tacit knowledge acquisition. The paper further proposes a habitus-driven individual development (HDID) Theoretical Model that can be of use to both academics and practitioners interested in fostering and developing processes of KS and individual development in knowledge-intensive organizations.
  16. Chen, H.: Knowledge-based document retrieval : framework and design (1992) 0.01
    0.008487149 = product of:
      0.06789719 = sum of:
        0.017106129 = weight(_text_:information in 5283) [ClassicSimilarity], result of:
          0.017106129 = score(doc=5283,freq=2.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.3103276 = fieldWeight in 5283, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=5283)
        0.050791066 = weight(_text_:retrieval in 5283) [ClassicSimilarity], result of:
          0.050791066 = score(doc=5283,freq=2.0), product of:
            0.09498371 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.031400457 = queryNorm
            0.5347345 = fieldWeight in 5283, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.125 = fieldNorm(doc=5283)
      0.125 = coord(2/16)
    
    Source
    Journal of information science. 18(1992), S.293-314
  17. Yang, M.; Kiang, M.; Chen, H.; Li, Y.: Artificial immune system for illicit content identification in social media (2012) 0.01
    0.008132801 = product of:
      0.04337494 = sum of:
        0.026127389 = weight(_text_:web in 4980) [ClassicSimilarity], result of:
          0.026127389 = score(doc=4980,freq=4.0), product of:
            0.10247572 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.031400457 = queryNorm
            0.25496176 = fieldWeight in 4980, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4980)
        0.00798859 = product of:
          0.01597718 = sum of:
            0.01597718 = weight(_text_:online in 4980) [ClassicSimilarity], result of:
              0.01597718 = score(doc=4980,freq=2.0), product of:
                0.09529729 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.031400457 = queryNorm
                0.16765618 = fieldWeight in 4980, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4980)
          0.5 = coord(1/2)
        0.009258964 = weight(_text_:information in 4980) [ClassicSimilarity], result of:
          0.009258964 = score(doc=4980,freq=6.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.16796975 = fieldWeight in 4980, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4980)
      0.1875 = coord(3/16)
    
    Abstract
    Social media is frequently used as a platform for the exchange of information and opinions as well as propaganda dissemination. But online content can be misused for the distribution of illicit information, such as violent postings in web forums. Illicit content is highly distributed in social media, while non-illicit content is unspecific and topically diverse. It is costly and time consuming to label a large amount of illicit content (positive examples) and non-illicit content (negative examples) to train classification systems. Nevertheless, it is relatively easy to obtain large volumes of unlabeled content in social media. In this article, an artificial immune system-based technique is presented to address the difficulties in the illicit content identification in social media. Inspired by the positive selection principle in the immune system, we designed a novel labeling heuristic based on partially supervised learning to extract high-quality positive and negative examples from unlabeled datasets. The empirical evaluation results from two large hate group web forums suggest that our proposed approach generally outperforms the benchmark techniques and exhibits more stable performance.
    Source
    Journal of the American Society for Information Science and Technology. 63(2012) no.2, S.256-269
  18. Chen, H.: Machine learning for information retrieval : neural networks, symbolic learning, and genetic algorithms (1994) 0.01
    0.007846749 = product of:
      0.062773995 = sum of:
        0.018331813 = weight(_text_:information in 2657) [ClassicSimilarity], result of:
          0.018331813 = score(doc=2657,freq=12.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.3325631 = fieldWeight in 2657, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2657)
        0.044442184 = weight(_text_:retrieval in 2657) [ClassicSimilarity], result of:
          0.044442184 = score(doc=2657,freq=8.0), product of:
            0.09498371 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.031400457 = queryNorm
            0.46789268 = fieldWeight in 2657, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2657)
      0.125 = coord(2/16)
    
    Abstract
    In the 1980s, knowledge-based techniques also made an impressive contribution to 'intelligent' information retrieval and indexing. More recently, researchers have turned to newer artificial intelligence based inductive learning techniques including neural networks, symbolic learning, and genetic algorithms grounded on diverse paradigms. These have provided great opportunities to enhance the capabilities of current information storage and retrieval systems. Provides an overview of these techniques and presents 3 popular methods: the connectionist Hopfield network; the symbolic ID3/ID5R; and evaluation based genetic algorithms in the context of information retrieval. The techniques are promising in their ability to analyze user queries, identify users' information needs, and suggest alternatives for search and can greatly complement the prevailing full text, keyword based, probabilistic, and knowledge based techniques
    Source
    Journal of the American Society for Information Science. 46(1995) no.3, S.194-216
  19. Chen, H.; Shankaranarayanan, G.; She, L.: ¬A machine learning approach to inductive query by examples : an experiment using relevance feedback, ID3, genetic algorithms, and simulated annealing (1998) 0.01
    0.00772925 = product of:
      0.061834 = sum of:
        0.019244393 = weight(_text_:information in 1148) [ClassicSimilarity], result of:
          0.019244393 = score(doc=1148,freq=18.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.34911853 = fieldWeight in 1148, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=1148)
        0.042589605 = weight(_text_:retrieval in 1148) [ClassicSimilarity], result of:
          0.042589605 = score(doc=1148,freq=10.0), product of:
            0.09498371 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.031400457 = queryNorm
            0.44838852 = fieldWeight in 1148, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=1148)
      0.125 = coord(2/16)
    
    Abstract
    Information retrieval using probabilistic techniques has attracted significant attention on the part of researchers in information and computer science over the past few decades. In the 1980s, knowledge-based techniques also made an impressive contribution to 'intelligent' information retrieval and indexing. More recently, information science researchers have tfurned to other newer inductive learning techniques including symbolic learning, genetic algorithms, and simulated annealing. These newer techniques, which are grounded in diverse paradigms, have provided great opportunities for researchers to enhance the information processing and retrieval capabilities of current information systems. In this article, we first provide an overview of these newer techniques and their use in information retrieval research. In order to femiliarize readers with the techniques, we present 3 promising methods: the symbolic ID3 algorithm, evolution-based genetic algorithms, and simulated annealing. We discuss their knowledge representations and algorithms in the unique context of information retrieval
    Source
    Journal of the American Society for Information Science. 49(1998) no.8, S.693-705
  20. Zheng, R.; Li, J.; Chen, H.; Huang, Z.: ¬A framework for authorship identification of online messages : writing-style features and classification techniques (2006) 0.01
    0.0072331186 = product of:
      0.038576633 = sum of:
        0.022595147 = product of:
          0.045190293 = sum of:
            0.045190293 = weight(_text_:online in 5276) [ClassicSimilarity], result of:
              0.045190293 = score(doc=5276,freq=16.0), product of:
                0.09529729 = queryWeight, product of:
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.031400457 = queryNorm
                0.47420335 = fieldWeight in 5276, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  3.0349014 = idf(docFreq=5778, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5276)
          0.5 = coord(1/2)
        0.005345665 = weight(_text_:information in 5276) [ClassicSimilarity], result of:
          0.005345665 = score(doc=5276,freq=2.0), product of:
            0.055122808 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.031400457 = queryNorm
            0.09697737 = fieldWeight in 5276, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5276)
        0.010635821 = product of:
          0.021271642 = sum of:
            0.021271642 = weight(_text_:22 in 5276) [ClassicSimilarity], result of:
              0.021271642 = score(doc=5276,freq=2.0), product of:
                0.10995905 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.031400457 = queryNorm
                0.19345059 = fieldWeight in 5276, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5276)
          0.5 = coord(1/2)
      0.1875 = coord(3/16)
    
    Abstract
    With the rapid proliferation of Internet technologies and applications, misuse of online messages for inappropriate or illegal purposes has become a major concern for society. The anonymous nature of online-message distribution makes identity tracing a critical problem. We developed a framework for authorship identification of online messages to address the identity-tracing problem. In this framework, four types of writing-style features (lexical, syntactic, structural, and content-specific features) are extracted and inductive learning algorithms are used to build feature-based classification models to identify authorship of online messages. To examine this framework, we conducted experiments on English and Chinese online-newsgroup messages. We compared the discriminating power of the four types of features and of three classification techniques: decision trees, backpropagation neural networks, and support vector machines. The experimental results showed that the proposed approach was able to identify authors of online messages with satisfactory accuracy of 70 to 95%. All four types of message features contributed to discriminating authors of online messages. Support vector machines outperformed the other two classification techniques in our experiments. The high performance we achieved for both the English and Chinese datasets showed the potential of this approach in a multiple-language context.
    Date
    22. 7.2006 16:14:37
    Source
    Journal of the American Society for Information Science and Technology. 57(2006) no.3, S.378-393

Types

  • a 61
  • el 1
  • More… Less…