Search (10 results, page 1 of 1)

  • × author_ss:"MacFarlane, A."
  1. MacFarlane, A.; Robertson, S.E.; McCann, J.A.: Parallel computing for passage retrieval (2004) 0.02
    0.01634812 = product of:
      0.09808872 = sum of:
        0.09808872 = sum of:
          0.05577461 = weight(_text_:methods in 5108) [ClassicSimilarity], result of:
            0.05577461 = score(doc=5108,freq=2.0), product of:
              0.15695344 = queryWeight, product of:
                4.0204134 = idf(docFreq=2156, maxDocs=44218)
                0.03903913 = queryNorm
              0.35535768 = fieldWeight in 5108, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.0204134 = idf(docFreq=2156, maxDocs=44218)
                0.0625 = fieldNorm(doc=5108)
          0.04231411 = weight(_text_:22 in 5108) [ClassicSimilarity], result of:
            0.04231411 = score(doc=5108,freq=2.0), product of:
              0.1367084 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03903913 = queryNorm
              0.30952093 = fieldWeight in 5108, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=5108)
      0.16666667 = coord(1/6)
    
    Abstract
    In this paper methods for both speeding up passage processing and examining more passages using parallel computers are explored. The number of passages processed are varied in order to examine the effect on retrieval effectiveness and efficiency. The particular algorithm applied has previously been used to good effect in Okapi experiments at TREC. This algorithm and the mechanism for applying parallel computing to speed up processing are described.
    Date
    20. 1.2007 18:30:22
  2. MacFarlane, A.; Tuson, A.: Local search : a guide for the information retrieval practitioner (2009) 0.01
    0.006037779 = product of:
      0.036226675 = sum of:
        0.036226675 = product of:
          0.07245335 = sum of:
            0.07245335 = weight(_text_:methods in 2457) [ClassicSimilarity], result of:
              0.07245335 = score(doc=2457,freq=6.0), product of:
                0.15695344 = queryWeight, product of:
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.03903913 = queryNorm
                0.4616232 = fieldWeight in 2457, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2457)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    There are a number of combinatorial optimisation problems in information retrieval in which the use of local search methods are worthwhile. The purpose of this paper is to show how local search can be used to solve some well known tasks in information retrieval (IR), how previous research in the field is piecemeal, bereft of a structure and methodologically flawed, and to suggest more rigorous ways of applying local search methods to solve IR problems. We provide a query based taxonomy for analysing the use of local search in IR tasks and an overview of issues such as fitness functions, statistical significance and test collections when conducting experiments on combinatorial optimisation problems. The paper gives a guide on the pitfalls and problems for IR practitioners who wish to use local search to solve their research issues, and gives practical advice on the use of such methods. The query based taxonomy is a novel structure which can be used by the IR practitioner in order to examine the use of local search in IR.
  3. MacFarlane, A.; McCann, J.A.; Robertson, S.E.: Parallel methods for the generation of partitioned inverted files (2005) 0.00
    0.004929826 = product of:
      0.029578956 = sum of:
        0.029578956 = product of:
          0.05915791 = sum of:
            0.05915791 = weight(_text_:methods in 651) [ClassicSimilarity], result of:
              0.05915791 = score(doc=651,freq=4.0), product of:
                0.15695344 = queryWeight, product of:
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.03903913 = queryNorm
                0.37691376 = fieldWeight in 651, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.046875 = fieldNorm(doc=651)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    Purpose - The generation of inverted indexes is one of the most computationally intensive activities for information retrieval systems: indexing large multi-gigabyte text databases can take many hours or even days to complete. We examine the generation of partitioned inverted files in order to speed up the process of indexing. Two types of index partitions are investigated: TermId and DocId. Design/methodology/approach - We use standard measures used in parallel computing such as speedup and efficiency to examine the computing results and also the space costs of our trial indexing experiments. Findings - The results from runs on both partitioning methods are compared and contrasted, concluding that DocId is the more efficient method. Practical implications - The practical implications are that the DocId partitioning method would in most circumstances be used for distributing inverted file data in a parallel computer, particularly if indexing speed is the primary consideration. Originality/value - The paper is of value to database administrators who manage large-scale text collections, and who need to use parallel computing to implement their text retrieval services.
  4. MacFarlane, A.; McCann, J.A.; Robertson, S.E.: Parallel methods for the update of partitioned inverted files (2007) 0.00
    0.004108188 = product of:
      0.024649128 = sum of:
        0.024649128 = product of:
          0.049298257 = sum of:
            0.049298257 = weight(_text_:methods in 819) [ClassicSimilarity], result of:
              0.049298257 = score(doc=819,freq=4.0), product of:
                0.15695344 = queryWeight, product of:
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.03903913 = queryNorm
                0.31409478 = fieldWeight in 819, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=819)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    Purpose - An issue that tends to be ignored in information retrieval is the issue of updating inverted files. This is largely because inverted files were devised to provide fast query service, and much work has been done with the emphasis strongly on queries. This paper aims to study the effect of using parallel methods for the update of inverted files in order to reduce costs, by looking at two types of partitioning for inverted files: document identifier and term identifier. Design/methodology/approach - Raw update service and update with query service are studied with these partitioning schemes using an incremental update strategy. The paper uses standard measures used in parallel computing such as speedup to examine the computing results and also the costs of reorganising indexes while servicing transactions. Findings - Empirical results show that for both transaction processing and index reorganisation the document identifier method is superior. However, there is evidence that the term identifier partitioning method could be useful in a concurrent transaction processing context. Practical implications - There is an increasing need to service updates, which is now becoming a requirement of inverted files (for dynamic collections such as the web), demonstrating that a shift in requirements of inverted file maintenance is needed from the past. Originality/value - The paper is of value to database administrators who manage large-scale and dynamic text collections, and who need to use parallel computing to implement their text retrieval services.
  5. Vakkari, P.; Jones, S.; MacFarlane, A.; Sormunen, E.: Query exhaustivity, relevance feedback and search success in automatic and interactive query expansion (2004) 0.00
    0.0034859132 = product of:
      0.020915478 = sum of:
        0.020915478 = product of:
          0.041830957 = sum of:
            0.041830957 = weight(_text_:methods in 4435) [ClassicSimilarity], result of:
              0.041830957 = score(doc=4435,freq=2.0), product of:
                0.15695344 = queryWeight, product of:
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.03903913 = queryNorm
                0.26651827 = fieldWeight in 4435, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4435)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    This study explored how the expression of search facets and relevance feedback (RF) by users was related to search success in interactive and automatic query expansion in the course of the search process. Search success was measured both in the number of relevant documents retrieved, whether identified by users or not. Research design consisted of 26 users searching for four TREC topics in Okapi IR system, half of the searchers using interactive and half automatic query expansion based on RF. The search logs were recorded, and the users filled in questionnaires for each topic concerning various features of searching. The results showed that the exhaustivity of the query was the most significant predictor of search success. Interactive expansion led to better search success than automatic expansion if all retrieved relevant items were counted, but there was no difference between the methods if only those items recognised relevant by users were observed. The analysis showed that the difference was facilitated by the liberal relevance criterion used in TREC not favouring highly relevant documents in evaluation.
  6. MacFarlane, A.: Evaluation of web search for the information practitioner (2007) 0.00
    0.0034859132 = product of:
      0.020915478 = sum of:
        0.020915478 = product of:
          0.041830957 = sum of:
            0.041830957 = weight(_text_:methods in 817) [ClassicSimilarity], result of:
              0.041830957 = score(doc=817,freq=2.0), product of:
                0.15695344 = queryWeight, product of:
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.03903913 = queryNorm
                0.26651827 = fieldWeight in 817, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.046875 = fieldNorm(doc=817)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    Purpose - The aim of the paper is to put forward a structured mechanism for web search evaluation. The paper seeks to point to useful scientific research and show how information practitioners can use these methods in evaluation of search on the web for their users. Design/methodology/approach - The paper puts forward an approach which utilizes traditional laboratory-based evaluation measures such as average precision/precision at N documents, augmented with diagnostic measures such as link broken, etc., which are used to show why precision measures are depressed as well as the quality of the search engines crawling mechanism. Findings - The paper shows how to use diagnostic measures in conjunction with precision in order to evaluate web search. Practical implications - The methodology presented in this paper will be useful to any information professional who regularly uses web search as part of their information seeking and needs to evaluate web search services. Originality/value - The paper argues that the use of diagnostic measures is essential in web search, as precision measures on their own do not allow a searcher to understand why search results differ between search engines.
  7. Inskip, C.; MacFarlane, A.; Rafferty, P.: Meaning, communication, music : towards a revised communication model (2008) 0.00
    0.0029049278 = product of:
      0.017429566 = sum of:
        0.017429566 = product of:
          0.034859132 = sum of:
            0.034859132 = weight(_text_:methods in 2347) [ClassicSimilarity], result of:
              0.034859132 = score(doc=2347,freq=2.0), product of:
                0.15695344 = queryWeight, product of:
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.03903913 = queryNorm
                0.22209854 = fieldWeight in 2347, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.0204134 = idf(docFreq=2156, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2347)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    Purpose - If an information retrieval system is going to be of value to the user then it must give meaning to the information which matches the meaning given to it by the user. The meaning given to music varies according to who is interpreting it - the author/composer, the performer, cataloguer or the listener - and this affects how music is organized and retrieved. This paper aims to examine the meaning of music, how meaning is communicated and suggests this may affect music retrieval. Design/methodology/approach - Musicology is used to define music and examine its functions leading to a discussion of how music has been organised and described. Various ways of establishing the meaning of music are reviewed, focussing on established musical analysis techniques. It is suggested that traditional methods are of limited use with digitised popular music. A discussion of semiotics and a review of semiotic analysis in western art music leads to a discussion of semiotics of popular music and examines ideas of Middleton, Stefani and Tagg. Findings - Agreeing that music exists when communication takes place, a discussion of selected communication models leads to the proposal of a revised version of Tagg's model, adjusting it to include listener feedback. Originality/value - The outcome of the analysis is a revised version of Tagg's communication model, adapted to reflect user feedback. It is suggested that this revised communication model reflects the way in which meaning is given to music.
  8. MacFarlane, A.; Missaoui, S.; Makri, S.; Gutierrez Lopez, M.: Sender vs. recipient-orientated information systems revisited (2022) 0.00
    0.0024861933 = product of:
      0.014917159 = sum of:
        0.014917159 = product of:
          0.029834319 = sum of:
            0.029834319 = weight(_text_:theory in 607) [ClassicSimilarity], result of:
              0.029834319 = score(doc=607,freq=2.0), product of:
                0.16234003 = queryWeight, product of:
                  4.1583924 = idf(docFreq=1878, maxDocs=44218)
                  0.03903913 = queryNorm
                0.18377672 = fieldWeight in 607, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.1583924 = idf(docFreq=1878, maxDocs=44218)
                  0.03125 = fieldNorm(doc=607)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    Purpose Belkin and Robertson (1976a) reflected on the ethical implications of theoretical research in information science and warned that there was potential for abuse of knowledge gained by undertaking such research and applying it to information systems. In particular, they identified the domains of advertising and political propaganda that posed particular problems. The purpose of this literature review is to revisit these ideas in the light of recent events in global information systems that demonstrate that their fears were justified. Design/methodology/approach The authors revisit the theory in information science that Belkin and Robertson used to build their argument, together with the discussion on ethics that resulted from this work in the late 1970s and early 1980s. The authors then review recent literature in the field of information systems, specifically information retrieval, social media and recommendation systems that highlight the problems identified by Belkin and Robertson. Findings Information science theories have been used in conjunction with empirical evidence gathered from user interactions that have been detrimental to both individuals and society. It is argued in the paper that the information science and systems communities should find ways to return control to the user wherever possible, and the ways to achieve this are considered. Research limitations/implications The ethical issues identified require a multidisciplinary approach with research in information science, computer science, information systems, business, sociology, psychology, journalism, government and politics, etc. required. This is too large a scope to deal with in a literature review, and we focus only on the design and implementation of information systems (Zimmer, 2008a) through an information science and information systems perspective. Practical implications The authors argue that information systems such as search technologies, social media applications and recommendation systems should be designed with the recipient of the information in mind (Paisley and Parker, 1965), not the sender of that information. Social implications Information systems designed ethically and with users in mind will go some way to addressing the ill effects typified by the problems for individuals and society evident in global information systems. Originality/value The authors synthesize the evidence from the literature to provide potential technological solutions to the ethical issues identified, with a set of recommendations to information systems designers and implementers.
  9. Inskip, C.; Butterworth, R.; MacFarlane, A.: ¬A study of the information needs of the users of a folk music library and the implications for the design of a digital library system (2008) 0.00
    0.0022238651 = product of:
      0.01334319 = sum of:
        0.01334319 = product of:
          0.02668638 = sum of:
            0.02668638 = weight(_text_:29 in 2053) [ClassicSimilarity], result of:
              0.02668638 = score(doc=2053,freq=2.0), product of:
                0.13732746 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03903913 = queryNorm
                0.19432661 = fieldWeight in 2053, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2053)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Date
    29. 7.2008 19:33:02
  10. Inskip, C.; MacFarlane, A.; Rafferty, P.: Organising music for movies (2010) 0.00
    0.0022238651 = product of:
      0.01334319 = sum of:
        0.01334319 = product of:
          0.02668638 = sum of:
            0.02668638 = weight(_text_:29 in 3941) [ClassicSimilarity], result of:
              0.02668638 = score(doc=3941,freq=2.0), product of:
                0.13732746 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03903913 = queryNorm
                0.19432661 = fieldWeight in 3941, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3941)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Date
    29. 8.2010 12:23:57