Search (15 results, page 1 of 1)

  • × author_ss:"Smeaton, A.F."
  • × year_i:[1990 TO 2000}
  1. Richardson, R.; Smeaton, A.F.; Murphy, J.: Using WordNet for conceptual distance measurement (1996) 0.03
    0.025941458 = product of:
      0.051882915 = sum of:
        0.051882915 = sum of:
          0.008202582 = weight(_text_:a in 6965) [ClassicSimilarity], result of:
            0.008202582 = score(doc=6965,freq=6.0), product of:
              0.053105544 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046056706 = queryNorm
              0.1544581 = fieldWeight in 6965, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0546875 = fieldNorm(doc=6965)
          0.043680333 = weight(_text_:22 in 6965) [ClassicSimilarity], result of:
            0.043680333 = score(doc=6965,freq=2.0), product of:
              0.16128273 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046056706 = queryNorm
              0.2708308 = fieldWeight in 6965, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=6965)
      0.5 = coord(1/2)
    
    Abstract
    Reports results of research to develop an information retrieval technique employing a conceptual distance measure between words and based on a large thesaurus. The techniques is specifically designed for data sharing in large scale autonomous distributed federated databases (FDBS). The prototype federated dictionary system, FEDDICT, stores information on the location of data sets within the FDBS and on semantic relationships exisitng between these data sets. WordNet is used and tested as the medium for bulding and operating FEDDICT
    Source
    Information retrieval: new systems and current research. Proceedings of the 16th Research Colloquium of the British Computer Society Information Retrieval Specialist Group, Drymen, Scotland, 22-23 Mar 94. Ed.: R. Leon
    Type
    a
  2. Kelledy, F.; Smeaton, A.F.: Signature files and beyond (1996) 0.02
    0.023258494 = product of:
      0.04651699 = sum of:
        0.04651699 = sum of:
          0.009076704 = weight(_text_:a in 6973) [ClassicSimilarity], result of:
            0.009076704 = score(doc=6973,freq=10.0), product of:
              0.053105544 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046056706 = queryNorm
              0.1709182 = fieldWeight in 6973, product of:
                3.1622777 = tf(freq=10.0), with freq of:
                  10.0 = termFreq=10.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=6973)
          0.037440285 = weight(_text_:22 in 6973) [ClassicSimilarity], result of:
            0.037440285 = score(doc=6973,freq=2.0), product of:
              0.16128273 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046056706 = queryNorm
              0.23214069 = fieldWeight in 6973, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=6973)
      0.5 = coord(1/2)
    
    Abstract
    Proposes that signature files be used as a viable alternative to other indexing strategies such as inverted files for searching through large volumes of text. Demonstrates through simulation, that search times can be further reduced by enhancing the basic signature file concept using deterministic partitioning algorithms which eliminate the need for an exhaustive search of the entire signature file. Reports research to evaluate the performance of some deterministic partitioning algorithms in a non simulated environment using 276 MB of raw newspaper text (taken from the Wall Street Journal) and real user queries. Presents a selection of results to illustrate trends and highlight important aspects of the performance of these methods under realistic rather than simulated operating conditions. As a result of the research reported here certain aspects of this approach to signature files are shown to be found wanting and require improvement. Suggests lines of future research on the partitioning of signature files
    Source
    Information retrieval: new systems and current research. Proceedings of the 16th Research Colloquium of the British Computer Society Information Retrieval Specialist Group, Drymen, Scotland, 22-23 Mar 94. Ed.: R. Leon
    Type
    a
  3. O'Donnell, R.; Smeaton, A.F.: ¬A linguistic approach to information retrieval (1996) 0.02
    0.022779368 = product of:
      0.045558736 = sum of:
        0.045558736 = sum of:
          0.008118451 = weight(_text_:a in 2575) [ClassicSimilarity], result of:
            0.008118451 = score(doc=2575,freq=8.0), product of:
              0.053105544 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046056706 = queryNorm
              0.15287387 = fieldWeight in 2575, product of:
                2.828427 = tf(freq=8.0), with freq of:
                  8.0 = termFreq=8.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=2575)
          0.037440285 = weight(_text_:22 in 2575) [ClassicSimilarity], result of:
            0.037440285 = score(doc=2575,freq=2.0), product of:
              0.16128273 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046056706 = queryNorm
              0.23214069 = fieldWeight in 2575, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2575)
      0.5 = coord(1/2)
    
    Abstract
    An important aspect of information retrieval systems is domain independence, where the subject of the information is not restricted to certain domains of knowledge. This should be able to represent any topic and although the text representation does not involve any semantic knowledge, lexical and syntactic analysis of the text allows the representation to remain domain independent. Reports research at Dublin City University, Ireland, which concentrates on the lexical and syntactic levels of natural language analysis and describes a domain independent automatic information retrieval system which accesses a very large database of newspaper text from the Wall Street Journal. The system represents the text in the form of syntax trees, and these trees are used in the matching process. Reports early results from the stuyd
    Source
    Information retrieval: new systems and current research. Proceedings of the 16th Research Colloquium of the British Computer Society Information Retrieval Specialist Group, Drymen, Scotland, 22-23 Mar 94. Ed.: R. Leon
    Type
    a
  4. Smeaton, A.F.: Retrieving information from hypertext : issues and problems (1991) 0.00
    0.003827074 = product of:
      0.007654148 = sum of:
        0.007654148 = product of:
          0.015308296 = sum of:
            0.015308296 = weight(_text_:a in 4278) [ClassicSimilarity], result of:
              0.015308296 = score(doc=4278,freq=16.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.28826174 = fieldWeight in 4278, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4278)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Hypertext uses a browsing rather than a searching strategy. Hypertext systems have found applications in a number of areas. They give users choice of information but this can prove a drawback. Examnines the effectiveness of hypertext as a way of retrieving information and reviews conventional information retrieval techniques. Considers previous attempts at combining information retrieval and hypertext and outlines a prototype systems developed to generate guided tours for users to direct them through hypertext to information they have requested. Discusses how adding this kind of itelligent guidance to a hypertext system would affect its usability as an information system
    Type
    a
  5. Smeaton, A.F.: Information retrieval and hypertext : competing technologies or complementary access methods (1992) 0.00
    0.00270615 = product of:
      0.0054123 = sum of:
        0.0054123 = product of:
          0.0108246 = sum of:
            0.0108246 = weight(_text_:a in 7503) [ClassicSimilarity], result of:
              0.0108246 = score(doc=7503,freq=2.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.20383182 = fieldWeight in 7503, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.125 = fieldNorm(doc=7503)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  6. Richardson, R.; Smeaton, A.F.: Automatic word sense disambiguation in a KBIR application (1995) 0.00
    0.00270615 = product of:
      0.0054123 = sum of:
        0.0054123 = product of:
          0.0108246 = sum of:
            0.0108246 = weight(_text_:a in 5796) [ClassicSimilarity], result of:
              0.0108246 = score(doc=5796,freq=8.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.20383182 = fieldWeight in 5796, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5796)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Discusses the implementation and design of an automatic word sense disambiguator. The semantic tagger is used in an overall Knowledge Based Information Retrieval (KBIR) system which uses a WordNet derived knowledge base (KB) and 2 independent semantic similarity estimators. The KB is used as a controlled vocabulary to represent documents and queries and the semantic similarity estimators are employed to determine the degree of relatedness between the KB representations
    Type
    a
  7. Smeaton, A.F.; Morrissey, P.J.: Experiments on the automatic construction of hypertext from texts (1995) 0.00
    0.0024857575 = product of:
      0.004971515 = sum of:
        0.004971515 = product of:
          0.00994303 = sum of:
            0.00994303 = weight(_text_:a in 7253) [ClassicSimilarity], result of:
              0.00994303 = score(doc=7253,freq=12.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.18723148 = fieldWeight in 7253, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=7253)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Describes an approach to semi-automatically generate a hypertext from linear texts, based on initially creatign nodes and composite nodes composed of 'mini-hypertexts'. Node-node similarity values are computed using standard information retrieval techniques and these similarity measures are then used to selectively create node-node links based on the strength of similarity between them. The process is a novel one because the link creation process also uses values from a dynamically computed metric which measures the topological compactness of the overall hypertext being generated. Describes experiments on generating a hypertext from a collection of 846 software product descriptions comprising 8,5 MBytes of text which yield some guidelines on how the process should be automated. This text to hypertext conversion method is put into the context of an overall hypertext authoring tool currently under development
    Type
    a
  8. Sheridan, P.; Smeaton, A.F.: ¬The application of morpho-syntactic language processing to effective phrase matching (1992) 0.00
    0.0023678814 = product of:
      0.0047357627 = sum of:
        0.0047357627 = product of:
          0.009471525 = sum of:
            0.009471525 = weight(_text_:a in 6575) [ClassicSimilarity], result of:
              0.009471525 = score(doc=6575,freq=2.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.17835285 = fieldWeight in 6575, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6575)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  9. Smeaton, A.F.: Natural language processing used in information retrieval tasks : an overview of achievements to date (1995) 0.00
    0.0023678814 = product of:
      0.0047357627 = sum of:
        0.0047357627 = product of:
          0.009471525 = sum of:
            0.009471525 = weight(_text_:a in 1265) [ClassicSimilarity], result of:
              0.009471525 = score(doc=1265,freq=2.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.17835285 = fieldWeight in 1265, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1265)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  10. Kelledy, F.; Smeaton, A.F.: Thresholding the postings lists in information retrieval : experiments on TREC data (1995) 0.00
    0.0023678814 = product of:
      0.0047357627 = sum of:
        0.0047357627 = product of:
          0.009471525 = sum of:
            0.009471525 = weight(_text_:a in 5804) [ClassicSimilarity], result of:
              0.009471525 = score(doc=5804,freq=8.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.17835285 = fieldWeight in 5804, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5804)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    A variety of methods for speeding up the response time of information retrieval processes have been put forward, one of which is the idea of thresholding. Thresholding relies on the data in information retrieval storage structures being organised to allow cut-off points to be used during processing. These cut-off points or thresholds are designed and ised to reduce the amount of information processed and to maintain the quality or minimise the degradation of response to a user's query. TREC is an annual series of benchmarking exercises to compare indexing and retrieval techniques. Reports experiments with a portion of the TREC data where features are introduced into the retrieval process to improve response time. These features improve response time while maintaining the same level of retrieval effectiveness
    Type
    a
  11. Smeaton, A.F.: Progress in the application of natural language processing to information retrieval tasks (1992) 0.00
    0.0020296127 = product of:
      0.0040592253 = sum of:
        0.0040592253 = product of:
          0.008118451 = sum of:
            0.008118451 = weight(_text_:a in 7080) [ClassicSimilarity], result of:
              0.008118451 = score(doc=7080,freq=2.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.15287387 = fieldWeight in 7080, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=7080)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  12. Kelledy, L.; Smeaton, A.F.: TREC-5 experiments at Dublin City University : Query space reduction, Spanish & character shape encoding (1997) 0.00
    0.0020296127 = product of:
      0.0040592253 = sum of:
        0.0040592253 = product of:
          0.008118451 = sum of:
            0.008118451 = weight(_text_:a in 3089) [ClassicSimilarity], result of:
              0.008118451 = score(doc=3089,freq=2.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.15287387 = fieldWeight in 3089, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3089)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  13. Smeaton, A.F.; Kelledy, L.; O'Donnell, R.: TREC-4 experiments at Dublin City University : thresholding posting lists, query expansion with WordNet and POS tagging of Spanish (1996) 0.00
    0.0016913437 = product of:
      0.0033826875 = sum of:
        0.0033826875 = product of:
          0.006765375 = sum of:
            0.006765375 = weight(_text_:a in 7000) [ClassicSimilarity], result of:
              0.006765375 = score(doc=7000,freq=2.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.12739488 = fieldWeight in 7000, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=7000)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a
  14. Smeaton, A.F.: Prospects for intelligent, language-based information retrieval (1991) 0.00
    0.001674345 = product of:
      0.00334869 = sum of:
        0.00334869 = product of:
          0.00669738 = sum of:
            0.00669738 = weight(_text_:a in 3700) [ClassicSimilarity], result of:
              0.00669738 = score(doc=3700,freq=4.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.12611452 = fieldWeight in 3700, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3700)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Current approaches to text retrieval based on indexing by words or index terms and on retrieving by specifying a Boolean combination of keywords are well known, as are their limitations. Statistical approaches to retrieval, as exemplified in commercial products like STATUS/IQ and Personal Librarian, are slightly better but still have their own weaknesses. Approaches to the indexing and retrieval of text based on techniques of automatic natural language processing (NLP) may soon start to realise their potential in terms of improving the quality and effectiveness of information retrieval. Examines some of the current attempts at using various NLP techniques in both the indexing and retrieval operations
    Type
    a
  15. Smeaton, A.F.; Harman, D.: ¬The TREC experiments and their impact on Europe (1997) 0.00
    0.001353075 = product of:
      0.00270615 = sum of:
        0.00270615 = product of:
          0.0054123 = sum of:
            0.0054123 = weight(_text_:a in 7702) [ClassicSimilarity], result of:
              0.0054123 = score(doc=7702,freq=2.0), product of:
                0.053105544 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046056706 = queryNorm
                0.10191591 = fieldWeight in 7702, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=7702)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Type
    a