Search (19 results, page 1 of 1)

  • × author_ss:"Wang, J."
  1. Shen, R.; Wang, J.; Fox, E.A.: ¬A Lightweight Protocol between Digital Libraries and Visualization Systems (2002) 0.03
    0.025946302 = product of:
      0.06486575 = sum of:
        0.01155891 = weight(_text_:a in 666) [ClassicSimilarity], result of:
          0.01155891 = score(doc=666,freq=4.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.2161963 = fieldWeight in 666, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.09375 = fieldNorm(doc=666)
        0.05330684 = product of:
          0.10661368 = sum of:
            0.10661368 = weight(_text_:22 in 666) [ClassicSimilarity], result of:
              0.10661368 = score(doc=666,freq=4.0), product of:
                0.16237405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046368346 = queryNorm
                0.6565931 = fieldWeight in 666, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=666)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Date
    22. 2.2003 17:25:39
    22. 2.2003 18:15:14
    Type
    a
  2. Hicks, D.; Wang, J.: Coverage and overlap of the new social sciences and humanities journal lists (2011) 0.02
    0.022870608 = product of:
      0.05717652 = sum of:
        0.0100103095 = weight(_text_:a in 4192) [ClassicSimilarity], result of:
          0.0100103095 = score(doc=4192,freq=12.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.18723148 = fieldWeight in 4192, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=4192)
        0.04716621 = sum of:
          0.009472587 = weight(_text_:information in 4192) [ClassicSimilarity], result of:
            0.009472587 = score(doc=4192,freq=2.0), product of:
              0.08139861 = queryWeight, product of:
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.046368346 = queryNorm
              0.116372846 = fieldWeight in 4192, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.046875 = fieldNorm(doc=4192)
          0.037693623 = weight(_text_:22 in 4192) [ClassicSimilarity], result of:
            0.037693623 = score(doc=4192,freq=2.0), product of:
              0.16237405 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046368346 = queryNorm
              0.23214069 = fieldWeight in 4192, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=4192)
      0.4 = coord(2/5)
    
    Abstract
    This is a study of coverage and overlap in second-generation social sciences and humanities journal lists, with attention paid to curation and the judgment of scholarliness. We identify four factors underpinning coverage shortfalls: journal language, country, publisher size, and age. Analyzing these factors turns our attention to the process of assessing a journal as scholarly, which is a necessary foundation for every list of scholarly journals. Although scholarliness should be a quality inherent in the journal, coverage falls short because groups assessing scholarliness have different perspectives on the social sciences and humanities literature. That the four factors shape perspectives on the literature points to a deeper problem of fragmentation within the scholarly community. We propose reducing this fragmentation as the best method to reduce coverage shortfalls.
    Date
    22. 1.2011 13:21:28
    Source
    Journal of the American Society for Information Science and Technology. 62(2011) no.2, S.284-294
    Type
    a
  3. Jiang, Z.; Gu, Q.; Yin, Y.; Wang, J.; Chen, D.: GRAW+ : a two-view graph propagation method with word coupling for readability assessment (2019) 0.02
    0.01905884 = product of:
      0.0476471 = sum of:
        0.008341924 = weight(_text_:a in 5218) [ClassicSimilarity], result of:
          0.008341924 = score(doc=5218,freq=12.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.15602624 = fieldWeight in 5218, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5218)
        0.039305177 = sum of:
          0.007893822 = weight(_text_:information in 5218) [ClassicSimilarity], result of:
            0.007893822 = score(doc=5218,freq=2.0), product of:
              0.08139861 = queryWeight, product of:
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.046368346 = queryNorm
              0.09697737 = fieldWeight in 5218, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5218)
          0.031411353 = weight(_text_:22 in 5218) [ClassicSimilarity], result of:
            0.031411353 = score(doc=5218,freq=2.0), product of:
              0.16237405 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046368346 = queryNorm
              0.19345059 = fieldWeight in 5218, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5218)
      0.4 = coord(2/5)
    
    Abstract
    Existing methods for readability assessment usually construct inductive classification models to assess the readability of singular text documents based on extracted features, which have been demonstrated to be effective. However, they rarely make use of the interrelationship among documents on readability, which can help increase the accuracy of readability assessment. In this article, we adopt a graph-based classification method to model and utilize the relationship among documents using the coupled bag-of-words model. We propose a word coupling method to build the coupled bag-of-words model by estimating the correlation between words on reading difficulty. In addition, we propose a two-view graph propagation method to make use of both the coupled bag-of-words model and the linguistic features. Our method employs a graph merging operation to combine graphs built according to different views, and improves the label propagation by incorporating the ordinal relation among reading levels. Experiments were conducted on both English and Chinese data sets, and the results demonstrate both effectiveness and potential of the method.
    Date
    15. 4.2019 13:46:22
    Source
    Journal of the Association for Information Science and Technology. 70(2019) no.5, S.433-447
    Type
    a
  4. He, R.; Wang, J.; Tian, J.; Chu, C.-T.; Mauney, B.; Perisic, I.: Session analysis of people search within a professional social network (2013) 0.02
    0.018446533 = product of:
      0.04611633 = sum of:
        0.0068111527 = weight(_text_:a in 743) [ClassicSimilarity], result of:
          0.0068111527 = score(doc=743,freq=8.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.12739488 = fieldWeight in 743, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=743)
        0.039305177 = sum of:
          0.007893822 = weight(_text_:information in 743) [ClassicSimilarity], result of:
            0.007893822 = score(doc=743,freq=2.0), product of:
              0.08139861 = queryWeight, product of:
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.046368346 = queryNorm
              0.09697737 = fieldWeight in 743, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.0390625 = fieldNorm(doc=743)
          0.031411353 = weight(_text_:22 in 743) [ClassicSimilarity], result of:
            0.031411353 = score(doc=743,freq=2.0), product of:
              0.16237405 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046368346 = queryNorm
              0.19345059 = fieldWeight in 743, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=743)
      0.4 = coord(2/5)
    
    Abstract
    We perform session analysis for our domain of people search within a professional social network. We find that the content-based method is appropriate to serve as a basis for the session identification in our domain. However, there remain some problems reported in previous research which degrade the identification performance (such as accuracy) of the content-based method. Therefore, in this article, we propose two important refinements to address these problems. We describe the underlying rationale of our refinements and then empirically show that the content-based method equipped with our refinements is able to achieve an excellent identification performance in our domain (such as 99.820% accuracy and 99.707% F-measure in our experiments). Next, because the time-based method has extremely low computation costs, which makes it suitable for many real-world applications, we investigate the feasibility of the time-based method in our domain by evaluating its identification performance based on our refined content-based method. Our experiments demonstrate that the performance of the time-based method is potentially acceptable to many real applications in our domain. Finally, we analyze several features of the identified sessions in our domain and compare them with the corresponding ones in general web search. The results illustrate the profession-oriented characteristics of our domain.
    Date
    19. 4.2013 20:31:22
    Source
    Journal of the American Society for Information Science and Technology. 64(2013) no.5, S.929-950
    Type
    a
  5. Wang, J.; Halffman, W.; Zhang, Y.H.: Sorting out journals : the proliferation of journal lists in China (2023) 0.02
    0.018081523 = product of:
      0.04520381 = sum of:
        0.005898632 = weight(_text_:a in 1055) [ClassicSimilarity], result of:
          0.005898632 = score(doc=1055,freq=6.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.11032722 = fieldWeight in 1055, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1055)
        0.039305177 = sum of:
          0.007893822 = weight(_text_:information in 1055) [ClassicSimilarity], result of:
            0.007893822 = score(doc=1055,freq=2.0), product of:
              0.08139861 = queryWeight, product of:
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.046368346 = queryNorm
              0.09697737 = fieldWeight in 1055, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.7554779 = idf(docFreq=20772, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1055)
          0.031411353 = weight(_text_:22 in 1055) [ClassicSimilarity], result of:
            0.031411353 = score(doc=1055,freq=2.0), product of:
              0.16237405 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046368346 = queryNorm
              0.19345059 = fieldWeight in 1055, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1055)
      0.4 = coord(2/5)
    
    Abstract
    Journal lists are instruments to categorize, compare, and assess research and scholarly publications. Our study investigates the remarkable proliferation of such journal lists in China, analyses their underlying values, quality criteria and ranking principles, and specifies how concerns specific to the Chinese research policy and publishing system inform these lists. Discouraged lists of "bad journals" reflect concerns over inferior research publications, but also the involved drain on public resources. Endorsed lists of "good journals" are based on criteria valued in research policy, reflecting the distinctive administrative logic of state-led Chinese research and publishing policy, ascribing worth to scientific journals for its specific national and institutional needs. In this regard, the criteria used for journal list construction are contextual and reflect the challenges of public resource allocation in a market-led publication system. Chinese journal lists therefore reflect research policy changes, such as a shift away from output-dominated research evaluation, the specific concerns about research misconduct, and balancing national research needs against international standards, resulting in distinctly Chinese quality criteria. However, contrasting concerns and inaccuracies lead to contradictions in the "qualify" and "disqualify" binary logic and demonstrate inherent tensions and limitations in journal lists as policy tools.
    Date
    22. 9.2023 16:39:23
    Source
    Journal of the Association for Information Science and Technology. 74(2023) no.10, S.1207-1228
    Type
    a
  6. Wang, J.; Reid, E.O.F.: Developing WWW information systems on the Internet (1996) 0.01
    0.009269844 = product of:
      0.02317461 = sum of:
        0.00770594 = weight(_text_:a in 604) [ClassicSimilarity], result of:
          0.00770594 = score(doc=604,freq=4.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.14413087 = fieldWeight in 604, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=604)
        0.01546867 = product of:
          0.03093734 = sum of:
            0.03093734 = weight(_text_:information in 604) [ClassicSimilarity], result of:
              0.03093734 = score(doc=604,freq=12.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.38007212 = fieldWeight in 604, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0625 = fieldNorm(doc=604)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Gives an overview of Web information system development. Discusses some basic concepts and technologies such as HTML, HTML FORM, CGI and Java, which are associated with developing WWW information systems. Further discusses the design and implementation of Virtual Travel Mart, a Web based end user oriented travel information system. Finally, addresses some issues in developing WWW information systems
    Source
    Microcomputers for information management. 13(1996) nos.3/4, S.237-252
    Type
    a
  7. Wang, J.; Oard, D.W.: Matching meaning for cross-language information retrieval (2012) 0.01
    0.008234787 = product of:
      0.020586967 = sum of:
        0.009535614 = weight(_text_:a in 7430) [ClassicSimilarity], result of:
          0.009535614 = score(doc=7430,freq=8.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.17835285 = fieldWeight in 7430, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=7430)
        0.011051352 = product of:
          0.022102704 = sum of:
            0.022102704 = weight(_text_:information in 7430) [ClassicSimilarity], result of:
              0.022102704 = score(doc=7430,freq=8.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.27153665 = fieldWeight in 7430, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=7430)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    This article describes a framework for cross-language information retrieval that efficiently leverages statistical estimation of translation probabilities. The framework provides a unified perspective into which some earlier work on techniques for cross-language information retrieval based on translation probabilities can be cast. Modeling synonymy and filtering translation probabilities using bidirectional evidence are shown to yield a balance between retrieval effectiveness and query-time (or indexing-time) efficiency that seems well suited large-scale applications. Evaluations with six test collections show consistent improvements over strong baselines.
    Source
    Information processing and management. 48(2012) no.4, S.631-653
    Type
    a
  8. Zhang, D.; Pee, L.G.; Pan, S.L.; Wang, J.: Information practices in data analytics for supporting public health surveillance (2024) 0.01
    0.007909955 = product of:
      0.019774888 = sum of:
        0.008173384 = weight(_text_:a in 1197) [ClassicSimilarity], result of:
          0.008173384 = score(doc=1197,freq=8.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.15287387 = fieldWeight in 1197, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=1197)
        0.011601503 = product of:
          0.023203006 = sum of:
            0.023203006 = weight(_text_:information in 1197) [ClassicSimilarity], result of:
              0.023203006 = score(doc=1197,freq=12.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.2850541 = fieldWeight in 1197, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1197)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Public health surveillance based on data analytics plays a crucial role in detecting and responding to public health crises, such as infectious disease outbreaks. Previous information science research on the topic has focused on developing analytical algorithms and visualization tools. This study seeks to extend the research by investigating information practices in data analytics for public health surveillance. Through a case study of how data analytics was conducted for surveilling Influenza A and COVID-19 outbreaks, both exploration information practices (i.e., probing, synthesizing, exchanging) and exploitation information practices (i.e., scavenging, adapting, outreaching) were identified and detailed. These findings enrich our empirical understanding of how data analytics can be implemented to support public health surveillance.
    Source
    Journal of the Association for Information Science and Technology. 75(2023) no.1, S.79-93
    Type
    a
  9. Oard, D.W.; He, D.; Wang, J.: User-assisted query translation for interactive cross-language information retrieval (2008) 0.01
    0.007058388 = product of:
      0.01764597 = sum of:
        0.008173384 = weight(_text_:a in 2030) [ClassicSimilarity], result of:
          0.008173384 = score(doc=2030,freq=8.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.15287387 = fieldWeight in 2030, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=2030)
        0.009472587 = product of:
          0.018945174 = sum of:
            0.018945174 = weight(_text_:information in 2030) [ClassicSimilarity], result of:
              0.018945174 = score(doc=2030,freq=8.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.23274569 = fieldWeight in 2030, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2030)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Interactive Cross-Language Information Retrieval (CLIR), a process in which searcher and system collaborate to find documents that satisfy an information need regardless of the language in which those documents are written, calls for designs in which synergies between searcher and system can be leveraged so that the strengths of one can cover weaknesses of the other. This paper describes an approach that employs user-assisted query translation to help searchers better understand the system's operation. Supporting interaction and interface designs are introduced, and results from three user studies are presented. The results indicate that experienced searchers presented with this new system evolve new search strategies that make effective use of the new capabilities, that they achieve retrieval effectiveness comparable to results obtained using fully automatic techniques, and that reported satisfaction with support for cross-language searching increased. The paper concludes with a description of a freely available interactive CLIR system that incorporates lessons learned from this research.
    Source
    Information processing and management. 44(2008) no.1, S.181-211
    Type
    a
  10. Wang, J.: ¬An extensive study on automated Dewey Decimal Classification (2009) 0.01
    0.005886516 = product of:
      0.01471629 = sum of:
        0.010769378 = weight(_text_:a in 3172) [ClassicSimilarity], result of:
          0.010769378 = score(doc=3172,freq=20.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.20142901 = fieldWeight in 3172, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3172)
        0.003946911 = product of:
          0.007893822 = sum of:
            0.007893822 = weight(_text_:information in 3172) [ClassicSimilarity], result of:
              0.007893822 = score(doc=3172,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.09697737 = fieldWeight in 3172, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3172)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    In this paper, we present a theoretical analysis and extensive experiments on the automated assignment of Dewey Decimal Classification (DDC) classes to bibliographic data with a supervised machine-learning approach. Library classification systems, such as the DDC, impose great obstacles on state-of-art text categorization (TC) technologies, including deep hierarchy, data sparseness, and skewed distribution. We first analyze statistically the document and category distributions over the DDC, and discuss the obstacles imposed by bibliographic corpora and library classification schemes on TC technology. To overcome these obstacles, we propose an innovative algorithm to reshape the DDC structure into a balanced virtual tree by balancing the category distribution and flattening the hierarchy. To improve the classification effectiveness to a level acceptable to real-world applications, we propose an interactive classification model that is able to predict a class of any depth within a limited number of user interactions. The experiments are conducted on a large bibliographic collection created by the Library of Congress within the science and technology domains over 10 years. With no more than three interactions, a classification accuracy of nearly 90% is achieved, thus providing a practical solution to the automatic bibliographic classification problem.
    Source
    Journal of the American Society for Information Science and Technology. 60(2009) no.11, S.2269-2286
    Type
    a
  11. Strzalkowski, T.; Guthrie, L.; Karlgren, J.; Leistensnider, J.; Lin, F.; Perez-Carballo, J.; Straszheim, T.; Wang, J.; Wilding, J.: Natural language information retrieval : TREC-5 report (1997) 0.01
    0.00588199 = product of:
      0.014704974 = sum of:
        0.0068111527 = weight(_text_:a in 3100) [ClassicSimilarity], result of:
          0.0068111527 = score(doc=3100,freq=2.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.12739488 = fieldWeight in 3100, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.078125 = fieldNorm(doc=3100)
        0.007893822 = product of:
          0.015787644 = sum of:
            0.015787644 = weight(_text_:information in 3100) [ClassicSimilarity], result of:
              0.015787644 = score(doc=3100,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.19395474 = fieldWeight in 3100, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3100)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Type
    a
  12. Lu, C.; Bu, Y.; Wang, J.; Ding, Y.; Torvik, V.; Schnaars, M.; Zhang, C.: Examining scientific writing styles from the perspective of linguistic complexity : a cross-level moderation model (2019) 0.01
    0.005549766 = product of:
      0.013874415 = sum of:
        0.009138121 = weight(_text_:a in 5219) [ClassicSimilarity], result of:
          0.009138121 = score(doc=5219,freq=10.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.1709182 = fieldWeight in 5219, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=5219)
        0.0047362936 = product of:
          0.009472587 = sum of:
            0.009472587 = weight(_text_:information in 5219) [ClassicSimilarity], result of:
              0.009472587 = score(doc=5219,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.116372846 = fieldWeight in 5219, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5219)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Publishing articles in high-impact English journals is difficult for scholars around the world, especially for non-native English-speaking scholars (NNESs), most of whom struggle with proficiency in English. To uncover the differences in English scientific writing between native English-speaking scholars (NESs) and NNESs, we collected a large-scale data set containing more than 150,000 full-text articles published in PLoS between 2006 and 2015. We divided these articles into three groups according to the ethnic backgrounds of the first and corresponding authors, obtained by Ethnea, and examined the scientific writing styles in English from a two-fold perspective of linguistic complexity: (a) syntactic complexity, including measurements of sentence length and sentence complexity; and (b) lexical complexity, including measurements of lexical diversity, lexical density, and lexical sophistication. The observations suggest marginal differences between groups in syntactical and lexical complexity.
    Source
    Journal of the Association for Information Science and Technology. 70(2019) no.5, S.462-475
    Type
    a
  13. Wang, J.; Guan, J.: ¬The analysis and evaluation of knowledge efficiency in research groups (2005) 0.01
    0.0051638708 = product of:
      0.012909677 = sum of:
        0.008173384 = weight(_text_:a in 4238) [ClassicSimilarity], result of:
          0.008173384 = score(doc=4238,freq=8.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.15287387 = fieldWeight in 4238, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=4238)
        0.0047362936 = product of:
          0.009472587 = sum of:
            0.009472587 = weight(_text_:information in 4238) [ClassicSimilarity], result of:
              0.009472587 = score(doc=4238,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.116372846 = fieldWeight in 4238, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4238)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    To study the knowledge creation process, we introduce a conceptual framework that captures the major goals and features of research organizations. The knowledge efficiency of research groups is then empirically studied. The budget of the projects and size of the research groups are inputs of the projects. To make the assessment more reasonable, two-dimensional indicators, including a domestic impact factor and an international impact factor, are jointly used to evaluate the research outputs for Chinese research groups through a Data Envelopment Analysis approach with preferences. Through comparisons of groups with the highest and lowest efficiency, we discover the critical factors influencing productivity and efficiency of these research groups based an the proposed framework. Finally, we provide some management suggestions for research groups to improve their knowledge creation efficiency.
    Source
    Journal of the American Society for Information Science and Technology. 56(2005) no.11, S.1217-1226
    Type
    a
  14. Wang, J.; Clements, M.; Yang, J.; Vries, A.P. de; Reinders, M.J.T.: Personalization of tagging systems (2010) 0.00
    0.004725861 = product of:
      0.011814652 = sum of:
        0.007078358 = weight(_text_:a in 4229) [ClassicSimilarity], result of:
          0.007078358 = score(doc=4229,freq=6.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.13239266 = fieldWeight in 4229, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=4229)
        0.0047362936 = product of:
          0.009472587 = sum of:
            0.009472587 = weight(_text_:information in 4229) [ClassicSimilarity], result of:
              0.009472587 = score(doc=4229,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.116372846 = fieldWeight in 4229, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4229)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Social media systems have encouraged end user participation in the Internet, for the purpose of storing and distributing Internet content, sharing opinions and maintaining relationships. Collaborative tagging allows users to annotate the resulting user-generated content, and enables effective retrieval of otherwise uncategorised data. However, compared to professional web content production, collaborative tagging systems face the challenge that end-users assign tags in an uncontrolled manner, resulting in unsystematic and inconsistent metadata. This paper introduces a framework for the personalization of social media systems. We pinpoint three tasks that would benefit from personalization: collaborative tagging, collaborative browsing and collaborative search. We propose a ranking model for each task that integrates the individual user's tagging history in the recommendation of tags and content, to align its suggestions to the individual user preferences. We demonstrate on two real data sets that for all three tasks, the personalized ranking should take into account both the user's own preference and the opinion of others.
    Source
    Information processing and management. 46(2010) no.1, S.58-70
    Type
    a
  15. Wang, J.: Automatic thesaurus development : term extraction from title metadata (2006) 0.00
    0.004303226 = product of:
      0.010758064 = sum of:
        0.0068111527 = weight(_text_:a in 5063) [ClassicSimilarity], result of:
          0.0068111527 = score(doc=5063,freq=8.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.12739488 = fieldWeight in 5063, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5063)
        0.003946911 = product of:
          0.007893822 = sum of:
            0.007893822 = weight(_text_:information in 5063) [ClassicSimilarity], result of:
              0.007893822 = score(doc=5063,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.09697737 = fieldWeight in 5063, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5063)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    The application of thesauri in networked environments is seriously hampered by the challenges of introducing new concepts and terminology into the formal controlled vocabulary, which is critical for enhancing its retrieval capability. The author describes an automated process of adding new terms to thesauri as entry vocabulary by analyzing the association between words/phrases extracted from bibliographic titles and subject descriptors in the metadata record (subject descriptors are terms assigned from controlled vocabularies of thesauri to describe the subjects of the objects [e.g., books, articles] represented by the metadata records). The investigated approach uses a corpus of metadata for scientific and technical (S&T) publications in which the titles contain substantive words for key topics. The three steps of the method are (a) extracting words and phrases from the title field of the metadata; (b) applying a method to identify and select the specific and meaningful keywords based on the associated controlled vocabulary terms from the thesaurus used to catalog the objects; and (c) inserting selected keywords into the thesaurus as new terms (most of them are in hierarchical relationships with the existing concepts), thereby updating the thesaurus with new terminology that is being used in the literature. The effectiveness of the method was demonstrated by an experiment with the Chinese Classification Thesaurus (CCT) and bibliographic data in China Machine-Readable Cataloging Record (MARC) format (CNMARC) provided by Peking University Library. This approach is equally effective in large-scale collections and in other languages.
    Source
    Journal of the American Society for Information Science and Technology. 57(2006) no.7, S.907-920
    Type
    a
  16. Qiu, J.; Zuo, M.; Wang, J.; Cai, C.: Knowledge order in an online knowledge community : group heterogeneity and two paths mediated by group interaction (2021) 0.00
    0.0039382176 = product of:
      0.009845544 = sum of:
        0.005898632 = weight(_text_:a in 310) [ClassicSimilarity], result of:
          0.005898632 = score(doc=310,freq=6.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.11032722 = fieldWeight in 310, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=310)
        0.003946911 = product of:
          0.007893822 = sum of:
            0.007893822 = weight(_text_:information in 310) [ClassicSimilarity], result of:
              0.007893822 = score(doc=310,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.09697737 = fieldWeight in 310, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=310)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Knowledge order in an online knowledge community (OKC) refers to a consensual version of collective knowledge in the creation of shared knowledge representation. Much previous research has been conducted in the context of the ordered structure of objective knowledge systems, but this does little to explain the microlevel order of knowledge after users contribute knowledge and achieve consensus through online interactions in OKC. Based on interactive team cognition theory and the stigmergy coordination mechanism, our research aims to investigate how knowledge and experience heterogeneity affect knowledge order effectiveness and efficiency through collaborative and communicative interaction. To test our hypotheses, we randomly collected the records of 250 articles from the English version of Wikipedia. Partial least squares structural equation modeling indicated that OKC favoring online collective knowledge order by limiting communicative interaction, as collaborative interaction is very effective in achieving knowledge order and in achieving it in a fast way. From our findings, scholars and practitioners are advised to pay attention to online knowledge order in the management and design of OKC.
    Source
    Journal of the Association for Information Science and Technology. 72(2021) no.8, S.1075-1091
    Type
    a
  17. Mao, J.; Xu, W.; Yang, Y.; Wang, J.; Yuille, A.L.: Explain images with multimodal recurrent neural networks (2014) 0.00
    0.002002062 = product of:
      0.0100103095 = sum of:
        0.0100103095 = weight(_text_:a in 1557) [ClassicSimilarity], result of:
          0.0100103095 = score(doc=1557,freq=12.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.18723148 = fieldWeight in 1557, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=1557)
      0.2 = coord(1/5)
    
    Abstract
    In this paper, we present a multimodal Recurrent Neural Network (m-RNN) model for generating novel sentence descriptions to explain the content of images. It directly models the probability distribution of generating a word given previous words and the image. Image descriptions are generated by sampling from this distribution. The model consists of two sub-networks: a deep recurrent neural network for sentences and a deep convolutional network for images. These two sub-networks interact with each other in a multimodal layer to form the whole m-RNN model. The effectiveness of our model is validated on three benchmark datasets (IAPR TC-12 [8], Flickr 8K [28], and Flickr 30K [13]). Our model outperforms the state-of-the-art generative method. In addition, the m-RNN model can be applied to retrieval tasks for retrieving images or sentences, and achieves significant performance improvement over the state-of-the-art methods which directly optimize the ranking objective function for retrieval.
    Type
    a
  18. Wang, J.: Chinese serials : history, characteristics, and cataloging considerations (2003) 0.00
    0.001651617 = product of:
      0.008258085 = sum of:
        0.008258085 = weight(_text_:a in 5496) [ClassicSimilarity], result of:
          0.008258085 = score(doc=5496,freq=6.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.1544581 = fieldWeight in 5496, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5496)
      0.2 = coord(1/5)
    
    Abstract
    Chinese serials are an indispensable component of American academic library collections that have Chinese language or studies programs. This special type of collection has not only attracted the interest of Chinese scholars, but has also been more in demand by university students, faculty and researchers in the related fields. Academic libraries, especially those outside East Asian collections, face multiple challenges in ensuring access to this unique material due to limited library budgets and cataloging staff. This article focuses on enhancing the understanding of Chinese serials and the challenges in processing and cataloging this type of material, including a brief history of Chinese serials, a description of their unique characteristics, and issues concerning cataloging practice.
    Type
    a
  19. Gauch, S.; Wang, J.: Corpus analysis for TREC 5 query expansion (1997) 0.00
    0.0016346768 = product of:
      0.008173384 = sum of:
        0.008173384 = weight(_text_:a in 5800) [ClassicSimilarity], result of:
          0.008173384 = score(doc=5800,freq=2.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.15287387 = fieldWeight in 5800, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.09375 = fieldNorm(doc=5800)
      0.2 = coord(1/5)
    
    Type
    a