Search (10 results, page 1 of 1)

  • × author_ss:"Melucci, M."
  1. Agosti, M.; Crestani, F.; Melucci, M.: Design and implementation of a tool for the automatic construction of hypertexts for information retrieval (1996) 0.00
    0.002814962 = product of:
      0.005629924 = sum of:
        0.005629924 = product of:
          0.011259848 = sum of:
            0.011259848 = weight(_text_:a in 5571) [ClassicSimilarity], result of:
              0.011259848 = score(doc=5571,freq=14.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.23593865 = fieldWeight in 5571, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5571)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Describes the design and implementation of TACHIR, a tool for the automatic construction of hypertexts for information retrieval. Through the use of an authoring methodology employing a set of well known information retrieval techniques, TACHIR automatically builds up a hypertext from a document collection. The structure of the hypertext reflects a 3 level conceptual model which enables navigation among documents, index terms, and concepts using automatically determined links. The hypertext is implemented using the HTML language. It can be distributed on different sites and different machines over the Internet, and it can be navigated using WWW interfaces
    Type
    a
  2. Melucci, M.: Passage retrieval : a probabilistic technique (1998) 0.00
    0.0026061484 = product of:
      0.0052122967 = sum of:
        0.0052122967 = product of:
          0.010424593 = sum of:
            0.010424593 = weight(_text_:a in 1150) [ClassicSimilarity], result of:
              0.010424593 = score(doc=1150,freq=12.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.21843673 = fieldWeight in 1150, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1150)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This paper presents a probabilistic technique to retrieve passages from texts having a large size or heterogeneous semantic content. The proposed technique is independent on any supporting auxiliary data, such as text structure, topic organization, or pre-defined text segments. A Bayesian framework implements the probabilistic technique. We carried out experiments to compare the probabilistique technique to one based on a text segmentation algorithm. In particular, the probabilistique technique is more effective than, or as effective as the one based on the text segmentation to retrieve small passages. Results show that passage size affects passage retrieval performance. Results do also suggest that text organization and query generality may have an impact on the difference in effectiveness between the two techniques
    Type
    a
  3. Melucci, M.; Orio, N.: Design, implementation, and evaluation of a methodology for automatic stemmer generation (2007) 0.00
    0.0026061484 = product of:
      0.0052122967 = sum of:
        0.0052122967 = product of:
          0.010424593 = sum of:
            0.010424593 = weight(_text_:a in 268) [ClassicSimilarity], result of:
              0.010424593 = score(doc=268,freq=12.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.21843673 = fieldWeight in 268, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=268)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The authors describe a statistical approach based on hidden Markov models (HMMs), for generating stemmers automatically. The proposed approach requires little effort to insert new languages in the system even if minimal linguistic knowledge is available. This is a key advantage especially for digital libraries, which are often developed for a specific institution or government because the program can manage a great amount of documents written in local languages. The evaluation described in the article shows that the stemmers implemented by means of HMMs are as effective as those based on linguistic rules.
    Type
    a
  4. Bacchin, M.; Ferro, N.; Melucci, M.: ¬A probabilistic model for stemmer generation (2005) 0.00
    0.0026061484 = product of:
      0.0052122967 = sum of:
        0.0052122967 = product of:
          0.010424593 = sum of:
            0.010424593 = weight(_text_:a in 1001) [ClassicSimilarity], result of:
              0.010424593 = score(doc=1001,freq=12.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.21843673 = fieldWeight in 1001, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1001)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    In this paper we will present a language-independent probabilistic model which can automatically generate stemmers. Stemmers can improve the retrieval effectiveness of information retrieval systems, however the designing and the implementation of stemmers requires a laborious amount of effort due to the fact that documents and queries are often written or spoken in several different languages. The probabilistic model proposed in this paper aims at the development of stemmers used for several languages. The proposed model describes the mutual reinforcement relationship between stems and derivations and then provides a probabilistic interpretation. A series of experiments shows that the stemmers generated by the probabilistic model are as effective as the ones based on linguistic knowledge.
    Type
    a
  5. Agosti, M.; Crestani, F.; Melucci, M.: On the use of information retrieval techniques for the automatic construction of hypertext (1997) 0.00
    0.0023790773 = product of:
      0.0047581545 = sum of:
        0.0047581545 = product of:
          0.009516309 = sum of:
            0.009516309 = weight(_text_:a in 150) [ClassicSimilarity], result of:
              0.009516309 = score(doc=150,freq=10.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.19940455 = fieldWeight in 150, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=150)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Introduces what automatic authoring of a hypertext for information retrieval means. The most difficult part of the automatic construction of a hypertext is the creation of links connecting documents or document fragments that are related. Becaus of this, to many researchers it seemed natural to use information retrieval techniques for this purpose, since information retrieval has always dealt with the construction of relationships between objects mutually relevant. Presents a survey of some of the attempts toward the automatic construction of hypertexts for information retrieval. Identifies and compares scope, advantages and limitations of different approaches. Points out the main and most successful current lines of research
    Footnote
    Contribution to a special issue on methods and tools for the automatic construction of hypertext
    Type
    a
  6. Buccio, E. Di; Melucci, M.; Moro, F.: Detecting verbose queries and improving information retrieval (2014) 0.00
    0.002279905 = product of:
      0.00455981 = sum of:
        0.00455981 = product of:
          0.00911962 = sum of:
            0.00911962 = weight(_text_:a in 2695) [ClassicSimilarity], result of:
              0.00911962 = score(doc=2695,freq=18.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.19109234 = fieldWeight in 2695, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2695)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Although most of the queries submitted to search engines are composed of a few keywords and have a length that ranges from three to six words, more than 15% of the total volume of the queries are verbose, introduce ambiguity and cause topic drifts. We consider verbosity a different property of queries from length since a verbose query is not necessarily long, it might be succinct and a short query might be verbose. This paper proposes a methodology to automatically detect verbose queries and conditionally modify queries. The methodology proposed in this paper exploits state-of-the-art classification algorithms, combines concepts from a large linguistic database and uses a topic gisting algorithm we designed for verbose query modification purposes. Our experimental results have been obtained using the TREC Robust track collection, thirty topics classified by difficulty degree, four queries per topic classified by verbosity and length, and human assessment of query verbosity. Our results suggest that the methodology for query modification conditioned to query verbosity detection and topic gisting is significantly effective and that query modification should be refined when topic difficulty and query verbosity are considered since these two properties interact and query verbosity is not straightforwardly related to query length.
    Type
    a
  7. Melucci, M.: Making digital libraries effective : automatic generation of links for similarity search across hyper-textbooks (2004) 0.00
    0.0022338415 = product of:
      0.004467683 = sum of:
        0.004467683 = product of:
          0.008935366 = sum of:
            0.008935366 = weight(_text_:a in 2226) [ClassicSimilarity], result of:
              0.008935366 = score(doc=2226,freq=12.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.18723148 = fieldWeight in 2226, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2226)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Textbooks are more available in electronic format now than in the past. Because textbooks are typically large, the end user needs effective tools to rapidly access information encapsulated in textbooks stored in digital libraries. Statistical similarity-based links among hypertextbooks are a means to provide those tools. In this paper, the design and the implementation of a tool that generates networks of links within and across hypertextbooks through a completely automatic and unsupervised procedure is described. The design is based an statistical techniques. The overall methodology is presented together with the results of a case study reached through a working prototype that shows that connecting hyper-textbooks is an efficient way to provide an effective retrieval capability.
    Type
    a
  8. Melucci, M.: Contextual search : a computational framework (2012) 0.00
    0.002149515 = product of:
      0.00429903 = sum of:
        0.00429903 = product of:
          0.00859806 = sum of:
            0.00859806 = weight(_text_:a in 4913) [ClassicSimilarity], result of:
              0.00859806 = score(doc=4913,freq=16.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.18016359 = fieldWeight in 4913, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4913)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The growing availability of data in electronic form, the expansion of the World Wide Web and the accessibility of computational methods for large-scale data processing have allowed researchers in Information Retrieval (IR) to design systems which can effectively and efficiently constrain search within the boundaries given by context, thus transforming classical search into contextual search. Contextual Search: A Computational Framework introduces contextual search within a computational framework based on contextual variables, contextual factors and statistical models. It describes how statistical models can process contextual variables to infer the contextual factors underlying the current search context. It also provides background to the subject by: placing it among other surveys on relevance, interaction, context, and behaviour; providing a description of the contextual variables used for implementing the statistical models which represent and predict relevance and contextual factors; and providing an overview of the evaluation methodologies and findings relevant to this subject. Contextual Search: A Computational Framework is a highly recommended read, both for beginners who are embarking on research in this area and as a useful reference for established IR researchers.
    Content
    Table of contents 1. Introduction 2. Query Intent 3. Personal Interest 4. Document Quality 5. Contextual Search Evaluation 6. Conclusions Acknowledgements References A. Implementations
  9. Agosti, M.; Melucci, M.: Information retrieval techniques for the automatic construction of hypertext (2000) 0.00
    0.0021279112 = product of:
      0.0042558224 = sum of:
        0.0042558224 = product of:
          0.008511645 = sum of:
            0.008511645 = weight(_text_:a in 4671) [ClassicSimilarity], result of:
              0.008511645 = score(doc=4671,freq=2.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.17835285 = fieldWeight in 4671, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4671)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  10. Melucci, M.; Orio, N.: Combining melody processing and information retrieval techniques : methodology, evaluation, and system implementation (2004) 0.00
    0.0010639556 = product of:
      0.0021279112 = sum of:
        0.0021279112 = product of:
          0.0042558224 = sum of:
            0.0042558224 = weight(_text_:a in 3087) [ClassicSimilarity], result of:
              0.0042558224 = score(doc=3087,freq=2.0), product of:
                0.04772363 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.041389145 = queryNorm
                0.089176424 = fieldWeight in 3087, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3087)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a