Search (6 results, page 1 of 1)

  • × author_ss:"Jiang, Y."
  1. Jiang, Y.; Bai, W.; Zhang, X.; Hu, J.: Wikipedia-based information content and semantic similarity computation (2017) 0.01
    0.010088081 = product of:
      0.040352322 = sum of:
        0.040352322 = product of:
          0.080704644 = sum of:
            0.080704644 = weight(_text_:methods in 2877) [ClassicSimilarity], result of:
              0.080704644 = score(doc=2877,freq=8.0), product of:
                0.18168657 = queryWeight, product of:
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.045191016 = queryNorm
                0.4441971 = fieldWeight in 2877, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2877)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    The Information Content (IC) of a concept is a fundamental dimension in computational linguistics. It enables a better understanding of concept's semantics. In the past, several approaches to compute IC of a concept have been proposed. However, there are some limitations such as the facts of relying on corpora availability, manual tagging, or predefined ontologies and fitting non-dynamic domains in the existing methods. Wikipedia provides a very large domain-independent encyclopedic repository and semantic network for computing IC of concepts with more coverage than usual ontologies. In this paper, we propose some novel methods to IC computation of a concept to solve the shortcomings of existing approaches. The presented methods focus on the IC computation of a concept (i.e., Wikipedia category) drawn from the Wikipedia category structure. We propose several new IC-based measures to compute the semantic similarity between concepts. The evaluation, based on several widely used benchmarks and a benchmark developed in ourselves, sustains the intuitions with respect to human judgments. Overall, some methods proposed in this paper have a good human correlation and constitute some effective ways of determining IC values for concepts and semantic similarity between concepts.
  2. Jiang, Y.; Zhang, X.; Tang, Y.; Nie, R.: Feature-based approaches to semantic similarity assessment of concepts using Wikipedia (2015) 0.01
    0.0071333502 = product of:
      0.028533401 = sum of:
        0.028533401 = product of:
          0.057066802 = sum of:
            0.057066802 = weight(_text_:methods in 2682) [ClassicSimilarity], result of:
              0.057066802 = score(doc=2682,freq=4.0), product of:
                0.18168657 = queryWeight, product of:
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.045191016 = queryNorm
                0.31409478 = fieldWeight in 2682, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2682)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    Semantic similarity assessment between concepts is an important task in many language related applications. In the past, several approaches to assess similarity by evaluating the knowledge modeled in an (or multiple) ontology (or ontologies) have been proposed. However, there are some limitations such as the facts of relying on predefined ontologies and fitting non-dynamic domains in the existing measures. Wikipedia provides a very large domain-independent encyclopedic repository and semantic network for computing semantic similarity of concepts with more coverage than usual ontologies. In this paper, we propose some novel feature based similarity assessment methods that are fully dependent on Wikipedia and can avoid most of the limitations and drawbacks introduced above. To implement similarity assessment based on feature by making use of Wikipedia, firstly a formal representation of Wikipedia concepts is presented. We then give a framework for feature based similarity based on the formal representation of Wikipedia concepts. Lastly, we investigate several feature based approaches to semantic similarity measures resulting from instantiations of the framework. The evaluation, based on several widely used benchmarks and a benchmark developed in ourselves, sustains the intuitions with respect to human judgements. Overall, several methods proposed in this paper have good human correlation and constitute some effective ways of determining similarity between Wikipedia concepts.
  3. Qu, R.; Fang, Y.; Bai, W.; Jiang, Y.: Computing semantic similarity based on novel models of semantic representation using Wikipedia (2018) 0.01
    0.0071333502 = product of:
      0.028533401 = sum of:
        0.028533401 = product of:
          0.057066802 = sum of:
            0.057066802 = weight(_text_:methods in 5052) [ClassicSimilarity], result of:
              0.057066802 = score(doc=5052,freq=4.0), product of:
                0.18168657 = queryWeight, product of:
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.045191016 = queryNorm
                0.31409478 = fieldWeight in 5052, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5052)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    Computing Semantic Similarity (SS) between concepts is one of the most critical issues in many domains such as Natural Language Processing and Artificial Intelligence. Over the years, several SS measurement methods have been proposed by exploiting different knowledge resources. Wikipedia provides a large domain-independent encyclopedic repository and a semantic network for computing SS between concepts. Traditional feature-based measures rely on linear combinations of different properties with two main limitations, the insufficient information and the loss of semantic information. In this paper, we propose several hybrid SS measurement approaches by using the Information Content (IC) and features of concepts, which avoid the limitations introduced above. Considering integrating discrete properties into one component, we present two models of semantic representation, called CORM and CARM. Then, we compute SS based on these models and take the IC of categories as a supplement of SS measurement. The evaluation, based on several widely used benchmarks and a benchmark developed by ourselves, sustains the intuitions with respect to human judgments. In summary, our approaches are more efficient in determining SS between concepts and have a better human correlation than previous methods such as Word2Vec and NASARI.
  4. Liu, Y.; Du, F.; Sun, J.; Silva, T.; Jiang, Y.; Zhu, T.: Identifying social roles using heterogeneous features in online social networks (2019) 0.01
    0.0060528484 = product of:
      0.024211394 = sum of:
        0.024211394 = product of:
          0.048422787 = sum of:
            0.048422787 = weight(_text_:methods in 5293) [ClassicSimilarity], result of:
              0.048422787 = score(doc=5293,freq=2.0), product of:
                0.18168657 = queryWeight, product of:
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.045191016 = queryNorm
                0.26651827 = fieldWeight in 5293, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5293)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    Role analysis plays an important role when exploring social media and knowledge-sharing platforms for designing marking strategies. However, current methods in role analysis have overlooked content generated by users (e.g., posts) in social media and hence focus more on user behavior analysis. The user-generated content is very important for characterizing users. In this paper, we propose a novel method which integrates both user behavior and posted content by users to identify roles in online social networks. The proposed method models a role as a joint distribution of Gaussian distribution and multinomial distribution, which represent user behavioral feature and content feature respectively. The proposed method can be used to determine the number of roles concerned automatically. The experimental results show that the proposed method can be used to identify various roles more effectively and to get more insights on such characteristics.
  5. Sun, J.; Zhu, M.; Jiang, Y.; Liu, Y.; Wu, L.L.: Hierarchical attention model for personalized tag recommendation : peer effects on information value perception (2021) 0.01
    0.0050440403 = product of:
      0.020176161 = sum of:
        0.020176161 = product of:
          0.040352322 = sum of:
            0.040352322 = weight(_text_:methods in 98) [ClassicSimilarity], result of:
              0.040352322 = score(doc=98,freq=2.0), product of:
                0.18168657 = queryWeight, product of:
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.045191016 = queryNorm
                0.22209854 = fieldWeight in 98, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=98)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    With the development of Web-based social networks, many personalized tag recommendation approaches based on multi-information have been proposed. Due to the differences in users' preferences, different users care about different kinds of information. In the meantime, different elements within each kind of information are differentially informative for user tagging behaviors. In this context, how to effectively integrate different elements and different information separately becomes a key part of tag recommendation. However, the existing methods ignore this key part. In order to address this problem, we propose a deep neural network for tag recommendation. Specifically, we model two important attentive aspects with a hierarchical attention model. For different user-item pairs, the bottom layered attention network models the influence of different elements on the features representation of the information while the top layered attention network models the attentive scores of different information. To verify the effectiveness of the proposed method, we conduct extensive experiments on two real-world data sets. The results show that using attention network and different kinds of information can significantly improve the performance of the recommendation model, and verify the effectiveness and superiority of our proposed model.
  6. Jiang, Y.; Meng, R.; Huang, Y.; Lu, W.; Liu, J.: Generating keyphrases for readers : a controllable keyphrase generation framework (2023) 0.00
    0.0038267244 = product of:
      0.015306897 = sum of:
        0.015306897 = product of:
          0.030613795 = sum of:
            0.030613795 = weight(_text_:22 in 1012) [ClassicSimilarity], result of:
              0.030613795 = score(doc=1012,freq=2.0), product of:
                0.15825124 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191016 = queryNorm
                0.19345059 = fieldWeight in 1012, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1012)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    22. 6.2023 14:55:20