Search (2 results, page 1 of 1)

  • × author_ss:"Boughanem, M."
  • × theme_ss:"Computerlinguistik"
  1. Lhadj, L.S.; Boughanem, M.; Amrouche, K.: Enhancing information retrieval through concept-based language modeling and semantic smoothing (2016) 0.02
    0.02187561 = product of:
      0.054689027 = sum of:
        0.013664948 = weight(_text_:m in 3221) [ClassicSimilarity], result of:
          0.013664948 = score(doc=3221,freq=2.0), product of:
            0.09940409 = queryWeight, product of:
              2.4884486 = idf(docFreq=9980, maxDocs=44218)
              0.03994621 = queryNorm
            0.13746867 = fieldWeight in 3221, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.4884486 = idf(docFreq=9980, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3221)
        0.04102408 = weight(_text_:n in 3221) [ClassicSimilarity], result of:
          0.04102408 = score(doc=3221,freq=2.0), product of:
            0.17223433 = queryWeight, product of:
              4.3116565 = idf(docFreq=1611, maxDocs=44218)
              0.03994621 = queryNorm
            0.23818761 = fieldWeight in 3221, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.3116565 = idf(docFreq=1611, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3221)
      0.4 = coord(2/5)
    
    Abstract
    Traditionally, many information retrieval models assume that terms occur in documents independently. Although these models have already shown good performance, the word independency assumption seems to be unrealistic from a natural language point of view, which considers that terms are related to each other. Therefore, such an assumption leads to two well-known problems in information retrieval (IR), namely, polysemy, or term mismatch, and synonymy. In language models, these issues have been addressed by considering dependencies such as bigrams, phrasal-concepts, or word relationships, but such models are estimated using simple n-grams or concept counting. In this paper, we address polysemy and synonymy mismatch with a concept-based language modeling approach that combines ontological concepts from external resources with frequently found collocations from the document collection. In addition, the concept-based model is enriched with subconcepts and semantic relationships through a semantic smoothing technique so as to perform semantic matching. Experiments carried out on TREC collections show that our model achieves significant improvements over a single word-based model and the Markov Random Field model (using a Markov classifier).
  2. Belbachir, F.; Boughanem, M.: Using language models to improve opinion detection (2018) 0.00
    0.0021863915 = product of:
      0.0109319575 = sum of:
        0.0109319575 = weight(_text_:m in 5044) [ClassicSimilarity], result of:
          0.0109319575 = score(doc=5044,freq=2.0), product of:
            0.09940409 = queryWeight, product of:
              2.4884486 = idf(docFreq=9980, maxDocs=44218)
              0.03994621 = queryNorm
            0.10997493 = fieldWeight in 5044, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.4884486 = idf(docFreq=9980, maxDocs=44218)
              0.03125 = fieldNorm(doc=5044)
      0.2 = coord(1/5)