Search (14 results, page 1 of 1)

  • × author_ss:"Wang, J."
  1. Wang, J.: Chinese serials : history, characteristics, and cataloging considerations (2003) 0.06
    0.062499635 = product of:
      0.09374945 = sum of:
        0.01867095 = weight(_text_:on in 5496) [ClassicSimilarity], result of:
          0.01867095 = score(doc=5496,freq=2.0), product of:
            0.109763056 = queryWeight, product of:
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.04990557 = queryNorm
            0.17010231 = fieldWeight in 5496, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5496)
        0.0750785 = product of:
          0.150157 = sum of:
            0.150157 = weight(_text_:demand in 5496) [ClassicSimilarity], result of:
              0.150157 = score(doc=5496,freq=2.0), product of:
                0.31127608 = queryWeight, product of:
                  6.237302 = idf(docFreq=234, maxDocs=44218)
                  0.04990557 = queryNorm
                0.48239172 = fieldWeight in 5496, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  6.237302 = idf(docFreq=234, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5496)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Chinese serials are an indispensable component of American academic library collections that have Chinese language or studies programs. This special type of collection has not only attracted the interest of Chinese scholars, but has also been more in demand by university students, faculty and researchers in the related fields. Academic libraries, especially those outside East Asian collections, face multiple challenges in ensuring access to this unique material due to limited library budgets and cataloging staff. This article focuses on enhancing the understanding of Chinese serials and the challenges in processing and cataloging this type of material, including a brief history of Chinese serials, a description of their unique characteristics, and issues concerning cataloging practice.
  2. Jiang, Z.; Gu, Q.; Yin, Y.; Wang, J.; Chen, D.: GRAW+ : a two-view graph propagation method with word coupling for readability assessment (2019) 0.03
    0.02905105 = product of:
      0.043576576 = sum of:
        0.026672786 = weight(_text_:on in 5218) [ClassicSimilarity], result of:
          0.026672786 = score(doc=5218,freq=8.0), product of:
            0.109763056 = queryWeight, product of:
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.04990557 = queryNorm
            0.24300331 = fieldWeight in 5218, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5218)
        0.01690379 = product of:
          0.03380758 = sum of:
            0.03380758 = weight(_text_:22 in 5218) [ClassicSimilarity], result of:
              0.03380758 = score(doc=5218,freq=2.0), product of:
                0.1747608 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04990557 = queryNorm
                0.19345059 = fieldWeight in 5218, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5218)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Existing methods for readability assessment usually construct inductive classification models to assess the readability of singular text documents based on extracted features, which have been demonstrated to be effective. However, they rarely make use of the interrelationship among documents on readability, which can help increase the accuracy of readability assessment. In this article, we adopt a graph-based classification method to model and utilize the relationship among documents using the coupled bag-of-words model. We propose a word coupling method to build the coupled bag-of-words model by estimating the correlation between words on reading difficulty. In addition, we propose a two-view graph propagation method to make use of both the coupled bag-of-words model and the linguistic features. Our method employs a graph merging operation to combine graphs built according to different views, and improves the label propagation by incorporating the ordinal relation among reading levels. Experiments were conducted on both English and Chinese data sets, and the results demonstrate both effectiveness and potential of the method.
    Date
    15. 4.2019 13:46:22
  3. Hicks, D.; Wang, J.: Coverage and overlap of the new social sciences and humanities journal lists (2011) 0.03
    0.028611436 = product of:
      0.042917155 = sum of:
        0.02263261 = weight(_text_:on in 4192) [ClassicSimilarity], result of:
          0.02263261 = score(doc=4192,freq=4.0), product of:
            0.109763056 = queryWeight, product of:
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.04990557 = queryNorm
            0.20619515 = fieldWeight in 4192, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.046875 = fieldNorm(doc=4192)
        0.020284547 = product of:
          0.040569093 = sum of:
            0.040569093 = weight(_text_:22 in 4192) [ClassicSimilarity], result of:
              0.040569093 = score(doc=4192,freq=2.0), product of:
                0.1747608 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04990557 = queryNorm
                0.23214069 = fieldWeight in 4192, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4192)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This is a study of coverage and overlap in second-generation social sciences and humanities journal lists, with attention paid to curation and the judgment of scholarliness. We identify four factors underpinning coverage shortfalls: journal language, country, publisher size, and age. Analyzing these factors turns our attention to the process of assessing a journal as scholarly, which is a necessary foundation for every list of scholarly journals. Although scholarliness should be a quality inherent in the journal, coverage falls short because groups assessing scholarliness have different perspectives on the social sciences and humanities literature. That the four factors shape perspectives on the literature points to a deeper problem of fragmentation within the scholarly community. We propose reducing this fragmentation as the best method to reduce coverage shortfalls.
    Date
    22. 1.2011 13:21:28
  4. Wang, J.; Halffman, W.; Zhang, Y.H.: Sorting out journals : the proliferation of journal lists in China (2023) 0.02
    0.023842867 = product of:
      0.0357643 = sum of:
        0.01886051 = weight(_text_:on in 1055) [ClassicSimilarity], result of:
          0.01886051 = score(doc=1055,freq=4.0), product of:
            0.109763056 = queryWeight, product of:
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.04990557 = queryNorm
            0.1718293 = fieldWeight in 1055, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1055)
        0.01690379 = product of:
          0.03380758 = sum of:
            0.03380758 = weight(_text_:22 in 1055) [ClassicSimilarity], result of:
              0.03380758 = score(doc=1055,freq=2.0), product of:
                0.1747608 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04990557 = queryNorm
                0.19345059 = fieldWeight in 1055, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1055)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Journal lists are instruments to categorize, compare, and assess research and scholarly publications. Our study investigates the remarkable proliferation of such journal lists in China, analyses their underlying values, quality criteria and ranking principles, and specifies how concerns specific to the Chinese research policy and publishing system inform these lists. Discouraged lists of "bad journals" reflect concerns over inferior research publications, but also the involved drain on public resources. Endorsed lists of "good journals" are based on criteria valued in research policy, reflecting the distinctive administrative logic of state-led Chinese research and publishing policy, ascribing worth to scientific journals for its specific national and institutional needs. In this regard, the criteria used for journal list construction are contextual and reflect the challenges of public resource allocation in a market-led publication system. Chinese journal lists therefore reflect research policy changes, such as a shift away from output-dominated research evaluation, the specific concerns about research misconduct, and balancing national research needs against international standards, resulting in distinctly Chinese quality criteria. However, contrasting concerns and inaccuracies lead to contradictions in the "qualify" and "disqualify" binary logic and demonstrate inherent tensions and limitations in journal lists as policy tools.
    Date
    22. 9.2023 16:39:23
  5. He, R.; Wang, J.; Tian, J.; Chu, C.-T.; Mauney, B.; Perisic, I.: Session analysis of people search within a professional social network (2013) 0.02
    0.020160122 = product of:
      0.030240182 = sum of:
        0.013336393 = weight(_text_:on in 743) [ClassicSimilarity], result of:
          0.013336393 = score(doc=743,freq=2.0), product of:
            0.109763056 = queryWeight, product of:
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.04990557 = queryNorm
            0.121501654 = fieldWeight in 743, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.0390625 = fieldNorm(doc=743)
        0.01690379 = product of:
          0.03380758 = sum of:
            0.03380758 = weight(_text_:22 in 743) [ClassicSimilarity], result of:
              0.03380758 = score(doc=743,freq=2.0), product of:
                0.1747608 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04990557 = queryNorm
                0.19345059 = fieldWeight in 743, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=743)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    We perform session analysis for our domain of people search within a professional social network. We find that the content-based method is appropriate to serve as a basis for the session identification in our domain. However, there remain some problems reported in previous research which degrade the identification performance (such as accuracy) of the content-based method. Therefore, in this article, we propose two important refinements to address these problems. We describe the underlying rationale of our refinements and then empirically show that the content-based method equipped with our refinements is able to achieve an excellent identification performance in our domain (such as 99.820% accuracy and 99.707% F-measure in our experiments). Next, because the time-based method has extremely low computation costs, which makes it suitable for many real-world applications, we investigate the feasibility of the time-based method in our domain by evaluating its identification performance based on our refined content-based method. Our experiments demonstrate that the performance of the time-based method is potentially acceptable to many real applications in our domain. Finally, we analyze several features of the identified sessions in our domain and compare them with the corresponding ones in general web search. The results illustrate the profession-oriented characteristics of our domain.
    Date
    19. 4.2013 20:31:22
  6. Shen, R.; Wang, J.; Fox, E.A.: ¬A Lightweight Protocol between Digital Libraries and Visualization Systems (2002) 0.02
    0.019124456 = product of:
      0.057373367 = sum of:
        0.057373367 = product of:
          0.114746734 = sum of:
            0.114746734 = weight(_text_:22 in 666) [ClassicSimilarity], result of:
              0.114746734 = score(doc=666,freq=4.0), product of:
                0.1747608 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04990557 = queryNorm
                0.6565931 = fieldWeight in 666, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=666)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 2.2003 17:25:39
    22. 2.2003 18:15:14
  7. Wang, J.: ¬An extensive study on automated Dewey Decimal Classification (2009) 0.01
    0.009940362 = product of:
      0.029821085 = sum of:
        0.029821085 = weight(_text_:on in 3172) [ClassicSimilarity], result of:
          0.029821085 = score(doc=3172,freq=10.0), product of:
            0.109763056 = queryWeight, product of:
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.04990557 = queryNorm
            0.271686 = fieldWeight in 3172, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3172)
      0.33333334 = coord(1/3)
    
    Abstract
    In this paper, we present a theoretical analysis and extensive experiments on the automated assignment of Dewey Decimal Classification (DDC) classes to bibliographic data with a supervised machine-learning approach. Library classification systems, such as the DDC, impose great obstacles on state-of-art text categorization (TC) technologies, including deep hierarchy, data sparseness, and skewed distribution. We first analyze statistically the document and category distributions over the DDC, and discuss the obstacles imposed by bibliographic corpora and library classification schemes on TC technology. To overcome these obstacles, we propose an innovative algorithm to reshape the DDC structure into a balanced virtual tree by balancing the category distribution and flattening the hierarchy. To improve the classification effectiveness to a level acceptable to real-world applications, we propose an interactive classification model that is able to predict a class of any depth within a limited number of user interactions. The experiments are conducted on a large bibliographic collection created by the Library of Congress within the science and technology domains over 10 years. With no more than three interactions, a classification accuracy of nearly 90% is achieved, thus providing a practical solution to the automatic bibliographic classification problem.
  8. Zhang, D.; Pee, L.G.; Pan, S.L.; Wang, J.: Information practices in data analytics for supporting public health surveillance (2024) 0.01
    0.009239726 = product of:
      0.027719175 = sum of:
        0.027719175 = weight(_text_:on in 1197) [ClassicSimilarity], result of:
          0.027719175 = score(doc=1197,freq=6.0), product of:
            0.109763056 = queryWeight, product of:
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.04990557 = queryNorm
            0.25253648 = fieldWeight in 1197, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.046875 = fieldNorm(doc=1197)
      0.33333334 = coord(1/3)
    
    Abstract
    Public health surveillance based on data analytics plays a crucial role in detecting and responding to public health crises, such as infectious disease outbreaks. Previous information science research on the topic has focused on developing analytical algorithms and visualization tools. This study seeks to extend the research by investigating information practices in data analytics for public health surveillance. Through a case study of how data analytics was conducted for surveilling Influenza A and COVID-19 outbreaks, both exploration information practices (i.e., probing, synthesizing, exchanging) and exploitation information practices (i.e., scavenging, adapting, outreaching) were identified and detailed. These findings enrich our empirical understanding of how data analytics can be implemented to support public health surveillance.
  9. Wang, J.; Oard, D.W.: Matching meaning for cross-language information retrieval (2012) 0.01
    0.008801571 = product of:
      0.026404712 = sum of:
        0.026404712 = weight(_text_:on in 7430) [ClassicSimilarity], result of:
          0.026404712 = score(doc=7430,freq=4.0), product of:
            0.109763056 = queryWeight, product of:
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.04990557 = queryNorm
            0.24056101 = fieldWeight in 7430, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.0546875 = fieldNorm(doc=7430)
      0.33333334 = coord(1/3)
    
    Abstract
    This article describes a framework for cross-language information retrieval that efficiently leverages statistical estimation of translation probabilities. The framework provides a unified perspective into which some earlier work on techniques for cross-language information retrieval based on translation probabilities can be cast. Modeling synonymy and filtering translation probabilities using bidirectional evidence are shown to yield a balance between retrieval effectiveness and query-time (or indexing-time) efficiency that seems well suited large-scale applications. Evaluations with six test collections show consistent improvements over strong baselines.
  10. Wang, J.; Reid, E.O.F.: Developing WWW information systems on the Internet (1996) 0.01
    0.007112743 = product of:
      0.021338228 = sum of:
        0.021338228 = weight(_text_:on in 604) [ClassicSimilarity], result of:
          0.021338228 = score(doc=604,freq=2.0), product of:
            0.109763056 = queryWeight, product of:
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.04990557 = queryNorm
            0.19440265 = fieldWeight in 604, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.0625 = fieldNorm(doc=604)
      0.33333334 = coord(1/3)
    
  11. Wang, J.; Clements, M.; Yang, J.; Vries, A.P. de; Reinders, M.J.T.: Personalization of tagging systems (2010) 0.01
    0.0053345575 = product of:
      0.016003672 = sum of:
        0.016003672 = weight(_text_:on in 4229) [ClassicSimilarity], result of:
          0.016003672 = score(doc=4229,freq=2.0), product of:
            0.109763056 = queryWeight, product of:
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.04990557 = queryNorm
            0.14580199 = fieldWeight in 4229, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.046875 = fieldNorm(doc=4229)
      0.33333334 = coord(1/3)
    
    Abstract
    Social media systems have encouraged end user participation in the Internet, for the purpose of storing and distributing Internet content, sharing opinions and maintaining relationships. Collaborative tagging allows users to annotate the resulting user-generated content, and enables effective retrieval of otherwise uncategorised data. However, compared to professional web content production, collaborative tagging systems face the challenge that end-users assign tags in an uncontrolled manner, resulting in unsystematic and inconsistent metadata. This paper introduces a framework for the personalization of social media systems. We pinpoint three tasks that would benefit from personalization: collaborative tagging, collaborative browsing and collaborative search. We propose a ranking model for each task that integrates the individual user's tagging history in the recommendation of tags and content, to align its suggestions to the individual user preferences. We demonstrate on two real data sets that for all three tasks, the personalized ranking should take into account both the user's own preference and the opinion of others.
  12. Mao, J.; Xu, W.; Yang, Y.; Wang, J.; Yuille, A.L.: Explain images with multimodal recurrent neural networks (2014) 0.01
    0.0053345575 = product of:
      0.016003672 = sum of:
        0.016003672 = weight(_text_:on in 1557) [ClassicSimilarity], result of:
          0.016003672 = score(doc=1557,freq=2.0), product of:
            0.109763056 = queryWeight, product of:
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.04990557 = queryNorm
            0.14580199 = fieldWeight in 1557, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.046875 = fieldNorm(doc=1557)
      0.33333334 = coord(1/3)
    
    Abstract
    In this paper, we present a multimodal Recurrent Neural Network (m-RNN) model for generating novel sentence descriptions to explain the content of images. It directly models the probability distribution of generating a word given previous words and the image. Image descriptions are generated by sampling from this distribution. The model consists of two sub-networks: a deep recurrent neural network for sentences and a deep convolutional network for images. These two sub-networks interact with each other in a multimodal layer to form the whole m-RNN model. The effectiveness of our model is validated on three benchmark datasets (IAPR TC-12 [8], Flickr 8K [28], and Flickr 30K [13]). Our model outperforms the state-of-the-art generative method. In addition, the m-RNN model can be applied to retrieval tasks for retrieving images or sentences, and achieves significant performance improvement over the state-of-the-art methods which directly optimize the ranking objective function for retrieval.
  13. Wang, J.: Automatic thesaurus development : term extraction from title metadata (2006) 0.00
    0.0044454644 = product of:
      0.013336393 = sum of:
        0.013336393 = weight(_text_:on in 5063) [ClassicSimilarity], result of:
          0.013336393 = score(doc=5063,freq=2.0), product of:
            0.109763056 = queryWeight, product of:
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.04990557 = queryNorm
            0.121501654 = fieldWeight in 5063, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5063)
      0.33333334 = coord(1/3)
    
    Abstract
    The application of thesauri in networked environments is seriously hampered by the challenges of introducing new concepts and terminology into the formal controlled vocabulary, which is critical for enhancing its retrieval capability. The author describes an automated process of adding new terms to thesauri as entry vocabulary by analyzing the association between words/phrases extracted from bibliographic titles and subject descriptors in the metadata record (subject descriptors are terms assigned from controlled vocabularies of thesauri to describe the subjects of the objects [e.g., books, articles] represented by the metadata records). The investigated approach uses a corpus of metadata for scientific and technical (S&T) publications in which the titles contain substantive words for key topics. The three steps of the method are (a) extracting words and phrases from the title field of the metadata; (b) applying a method to identify and select the specific and meaningful keywords based on the associated controlled vocabulary terms from the thesaurus used to catalog the objects; and (c) inserting selected keywords into the thesaurus as new terms (most of them are in hierarchical relationships with the existing concepts), thereby updating the thesaurus with new terminology that is being used in the literature. The effectiveness of the method was demonstrated by an experiment with the Chinese Classification Thesaurus (CCT) and bibliographic data in China Machine-Readable Cataloging Record (MARC) format (CNMARC) provided by Peking University Library. This approach is equally effective in large-scale collections and in other languages.
  14. Qiu, J.; Zuo, M.; Wang, J.; Cai, C.: Knowledge order in an online knowledge community : group heterogeneity and two paths mediated by group interaction (2021) 0.00
    0.0044454644 = product of:
      0.013336393 = sum of:
        0.013336393 = weight(_text_:on in 310) [ClassicSimilarity], result of:
          0.013336393 = score(doc=310,freq=2.0), product of:
            0.109763056 = queryWeight, product of:
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.04990557 = queryNorm
            0.121501654 = fieldWeight in 310, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.199415 = idf(docFreq=13325, maxDocs=44218)
              0.0390625 = fieldNorm(doc=310)
      0.33333334 = coord(1/3)
    
    Abstract
    Knowledge order in an online knowledge community (OKC) refers to a consensual version of collective knowledge in the creation of shared knowledge representation. Much previous research has been conducted in the context of the ordered structure of objective knowledge systems, but this does little to explain the microlevel order of knowledge after users contribute knowledge and achieve consensus through online interactions in OKC. Based on interactive team cognition theory and the stigmergy coordination mechanism, our research aims to investigate how knowledge and experience heterogeneity affect knowledge order effectiveness and efficiency through collaborative and communicative interaction. To test our hypotheses, we randomly collected the records of 250 articles from the English version of Wikipedia. Partial least squares structural equation modeling indicated that OKC favoring online collective knowledge order by limiting communicative interaction, as collaborative interaction is very effective in achieving knowledge order and in achieving it in a fast way. From our findings, scholars and practitioners are advised to pay attention to online knowledge order in the management and design of OKC.