Search (12 results, page 1 of 1)

  • × author_ss:"MacFarlane, A."
  1. Inskip, C.; Butterworth, R.; MacFarlane, A.: ¬A study of the information needs of the users of a folk music library and the implications for the design of a digital library system (2008) 0.04
    0.04400172 = product of:
      0.0733362 = sum of:
        0.040348392 = weight(_text_:context in 2053) [ClassicSimilarity], result of:
          0.040348392 = score(doc=2053,freq=2.0), product of:
            0.17622331 = queryWeight, product of:
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.04251826 = queryNorm
            0.22896172 = fieldWeight in 2053, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2053)
        0.023299592 = weight(_text_:system in 2053) [ClassicSimilarity], result of:
          0.023299592 = score(doc=2053,freq=2.0), product of:
            0.13391352 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.04251826 = queryNorm
            0.17398985 = fieldWeight in 2053, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2053)
        0.009688215 = product of:
          0.029064644 = sum of:
            0.029064644 = weight(_text_:29 in 2053) [ClassicSimilarity], result of:
              0.029064644 = score(doc=2053,freq=2.0), product of:
                0.14956595 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04251826 = queryNorm
                0.19432661 = fieldWeight in 2053, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2053)
          0.33333334 = coord(1/3)
      0.6 = coord(3/5)
    
    Date
    29. 7.2008 19:33:02
    Footnote
    Beitrag eines Themenschwerpunktes "Digital libraries in the context of users' broader activities"
  2. Lu, W.; MacFarlane, A.; Venuti, F.: Okapi-based XML indexing (2009) 0.03
    0.0340826 = product of:
      0.085206494 = sum of:
        0.044850416 = weight(_text_:index in 3629) [ClassicSimilarity], result of:
          0.044850416 = score(doc=3629,freq=2.0), product of:
            0.18579477 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.04251826 = queryNorm
            0.24139762 = fieldWeight in 3629, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3629)
        0.04035608 = weight(_text_:system in 3629) [ClassicSimilarity], result of:
          0.04035608 = score(doc=3629,freq=6.0), product of:
            0.13391352 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.04251826 = queryNorm
            0.30135927 = fieldWeight in 3629, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3629)
      0.4 = coord(2/5)
    
    Abstract
    Purpose - Being an important data exchange and information storage standard, XML has generated a great deal of interest and particular attention has been paid to the issue of XML indexing. Clear use cases for structured search in XML have been established. However, most of the research in the area is either based on relational database systems or specialized semi-structured data management systems. This paper aims to propose a method for XML indexing based on the information retrieval (IR) system Okapi. Design/methodology/approach - First, the paper reviews the structure of inverted files and gives an overview of the issues of why this indexing mechanism cannot properly support XML retrieval, using the underlying data structures of Okapi as an example. Then the paper explores a revised method implemented on Okapi using path indexing structures. The paper evaluates these index structures through the metrics of indexing run time, path search run time and space costs using the INEX and Reuters RVC1 collections. Findings - Initial results on the INEX collections show that there is a substantial overhead in space costs for the method, but this increase does not affect run time adversely. Indexing results on differing sized Reuters RVC1 sub-collections show that the increase in space costs with increasing the size of a collection is significant, but in terms of run time the increase is linear. Path search results show sub-millisecond run times, demonstrating minimal overhead for XML search. Practical implications - Overall, the results show the method implemented to support XML search in a traditional IR system such as Okapi is viable. Originality/value - The paper provides useful information on a method for XML indexing based on the IR system Okapi.
  3. MacFarlane, A.; McCann, J.A.; Robertson, S.E.: Parallel methods for the update of partitioned inverted files (2007) 0.03
    0.034079522 = product of:
      0.085198805 = sum of:
        0.040348392 = weight(_text_:context in 819) [ClassicSimilarity], result of:
          0.040348392 = score(doc=819,freq=2.0), product of:
            0.17622331 = queryWeight, product of:
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.04251826 = queryNorm
            0.22896172 = fieldWeight in 819, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.0390625 = fieldNorm(doc=819)
        0.044850416 = weight(_text_:index in 819) [ClassicSimilarity], result of:
          0.044850416 = score(doc=819,freq=2.0), product of:
            0.18579477 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.04251826 = queryNorm
            0.24139762 = fieldWeight in 819, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.0390625 = fieldNorm(doc=819)
      0.4 = coord(2/5)
    
    Abstract
    Purpose - An issue that tends to be ignored in information retrieval is the issue of updating inverted files. This is largely because inverted files were devised to provide fast query service, and much work has been done with the emphasis strongly on queries. This paper aims to study the effect of using parallel methods for the update of inverted files in order to reduce costs, by looking at two types of partitioning for inverted files: document identifier and term identifier. Design/methodology/approach - Raw update service and update with query service are studied with these partitioning schemes using an incremental update strategy. The paper uses standard measures used in parallel computing such as speedup to examine the computing results and also the costs of reorganising indexes while servicing transactions. Findings - Empirical results show that for both transaction processing and index reorganisation the document identifier method is superior. However, there is evidence that the term identifier partitioning method could be useful in a concurrent transaction processing context. Practical implications - There is an increasing need to service updates, which is now becoming a requirement of inverted files (for dynamic collections such as the web), demonstrating that a shift in requirements of inverted file maintenance is needed from the past. Originality/value - The paper is of value to database administrators who manage large-scale and dynamic text collections, and who need to use parallel computing to implement their text retrieval services.
  4. MacFarlane, A.; McCann, J.A.; Robertson, S.E.: Parallel methods for the generation of partitioned inverted files (2005) 0.01
    0.0107641 = product of:
      0.0538205 = sum of:
        0.0538205 = weight(_text_:index in 651) [ClassicSimilarity], result of:
          0.0538205 = score(doc=651,freq=2.0), product of:
            0.18579477 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.04251826 = queryNorm
            0.28967714 = fieldWeight in 651, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.046875 = fieldNorm(doc=651)
      0.2 = coord(1/5)
    
    Abstract
    Purpose - The generation of inverted indexes is one of the most computationally intensive activities for information retrieval systems: indexing large multi-gigabyte text databases can take many hours or even days to complete. We examine the generation of partitioned inverted files in order to speed up the process of indexing. Two types of index partitions are investigated: TermId and DocId. Design/methodology/approach - We use standard measures used in parallel computing such as speedup and efficiency to examine the computing results and also the space costs of our trial indexing experiments. Findings - The results from runs on both partitioning methods are compared and contrasted, concluding that DocId is the more efficient method. Practical implications - The practical implications are that the DocId partitioning method would in most circumstances be used for distributing inverted file data in a parallel computer, particularly if indexing speed is the primary consideration. Originality/value - The paper is of value to database administrators who manage large-scale text collections, and who need to use parallel computing to implement their text retrieval services.
  5. MacFarlane, A.; Missaoui, S.; Frankowska-Takhari, S.: On machine learning and knowledge organization in multimedia information retrieval (2020) 0.01
    0.008970084 = product of:
      0.044850416 = sum of:
        0.044850416 = weight(_text_:index in 5732) [ClassicSimilarity], result of:
          0.044850416 = score(doc=5732,freq=2.0), product of:
            0.18579477 = queryWeight, product of:
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.04251826 = queryNorm
            0.24139762 = fieldWeight in 5732, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.369764 = idf(docFreq=1520, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5732)
      0.2 = coord(1/5)
    
    Abstract
    Recent technological developments have increased the use of machine learning to solve many problems, including many in information retrieval. Multimedia information retrieval as a problem represents a significant challenge to machine learning as a technological solution, but some problems can still be addressed by using appropriate AI techniques. We review the technological developments and provide a perspective on the use of machine learning in conjunction with knowledge organization to address multimedia IR needs. The semantic gap in multimedia IR remains a significant problem in the field, and solutions to them are many years off. However, new technological developments allow the use of knowledge organization and machine learning in multimedia search systems and services. Specifically, we argue that, the improvement of detection of some classes of lowlevel features in images music and video can be used in conjunction with knowledge organization to tag or label multimedia content for better retrieval performance. We provide an overview of the use of knowledge organization schemes in machine learning and make recommendations to information professionals on the use of this technology with knowledge organization techniques to solve multimedia IR problems. We introduce a five-step process model that extracts features from multimedia objects (Step 1) from both knowledge organization (Step 1a) and machine learning (Step 1b), merging them together (Step 2) to create an index of those multimedia objects (Step 3). We also overview further steps in creating an application to utilize the multimedia objects (Step 4) and maintaining and updating the database of features on those objects (Step 5).
  6. Konkova, E.; MacFarlane, A.; Göker, A.: Analysing creative image search information needs (2016) 0.01
    0.008069678 = product of:
      0.040348392 = sum of:
        0.040348392 = weight(_text_:context in 2833) [ClassicSimilarity], result of:
          0.040348392 = score(doc=2833,freq=2.0), product of:
            0.17622331 = queryWeight, product of:
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.04251826 = queryNorm
            0.22896172 = fieldWeight in 2833, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2833)
      0.2 = coord(1/5)
    
    Abstract
    Creative professionals in advertising, marketing, design and journalism search for images to visually represent a concept for their project. The main purpose of this paper is to present search facets derived from an analysis of documents known as briefs, which are widely used in creative industries as requirement documents describing information needs. The briefs specify the type of image required, such as the content and context of use for the image and represent the topic from which the searcher builds an image query. We take three main sources-user image search behaviour, briefs, and image search engine search facets-to examine the search facets for image searching in order to examine the following research question-are search facet schemes for image search engines sufficient for user needs, or is revision needed? We found that there are three main classes of user search facet, which include business, contextual and image related information. The key argument in the paper is that the facet "keyword/tag" is ambiguous and does not support user needs for more generic descriptions to broaden search or specific descriptions to narrow their search-we suggest that a more detailed search facet scheme would be appropriate.
  7. Berget, G.; MacFarlane, A.: What Is known about the impact of impairments on information seeking and searching? (2020) 0.01
    0.008069678 = product of:
      0.040348392 = sum of:
        0.040348392 = weight(_text_:context in 5817) [ClassicSimilarity], result of:
          0.040348392 = score(doc=5817,freq=2.0), product of:
            0.17622331 = queryWeight, product of:
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.04251826 = queryNorm
            0.22896172 = fieldWeight in 5817, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.14465 = idf(docFreq=1904, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5817)
      0.2 = coord(1/5)
    
    Abstract
    Information seeking and access are essential for users in all walks of life, from addressing personal needs such as finding flights to locating information needed to complete work tasks. Over the past decade or so, the general needs of people with impairments have increasingly been recognized as something to be addressed, an issue embedded both in international treaties and in state legislation. The same tendency can be found in research, where a growing number of user studies including people with impairments have been conducted. The purpose of these studies is typically to uncover potential barriers for access to information, especially in the context of inaccessible search user interfaces. This literature review provides an overview of research on the information seeking and searching of users with impairments. The aim is to provide an overview to both researchers and practitioners who work with any of the user groups identified. Some diagnoses are relatively well represented in the literature (for instance, visual impairment), but there is very little work in other areas (for instance, autism) and in some cases no work at all (for instance, aphasia). Gaps are identified in the research, and suggestions are made regarding areas where further research is needed.
  8. Vakkari, P.; Jones, S.; MacFarlane, A.; Sormunen, E.: Query exhaustivity, relevance feedback and search success in automatic and interactive query expansion (2004) 0.01
    0.0055919024 = product of:
      0.027959513 = sum of:
        0.027959513 = weight(_text_:system in 4435) [ClassicSimilarity], result of:
          0.027959513 = score(doc=4435,freq=2.0), product of:
            0.13391352 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.04251826 = queryNorm
            0.20878783 = fieldWeight in 4435, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.046875 = fieldNorm(doc=4435)
      0.2 = coord(1/5)
    
    Abstract
    This study explored how the expression of search facets and relevance feedback (RF) by users was related to search success in interactive and automatic query expansion in the course of the search process. Search success was measured both in the number of relevant documents retrieved, whether identified by users or not. Research design consisted of 26 users searching for four TREC topics in Okapi IR system, half of the searchers using interactive and half automatic query expansion based on RF. The search logs were recorded, and the users filled in questionnaires for each topic concerning various features of searching. The results showed that the exhaustivity of the query was the most significant predictor of search success. Interactive expansion led to better search success than automatic expansion if all retrieved relevant items were counted, but there was no difference between the methods if only those items recognised relevant by users were observed. The analysis showed that the difference was facilitated by the liberal relevance criterion used in TREC not favouring highly relevant documents in evaluation.
  9. Inskip, C.; MacFarlane, A.; Rafferty, P.: Meaning, communication, music : towards a revised communication model (2008) 0.00
    0.0046599186 = product of:
      0.023299592 = sum of:
        0.023299592 = weight(_text_:system in 2347) [ClassicSimilarity], result of:
          0.023299592 = score(doc=2347,freq=2.0), product of:
            0.13391352 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.04251826 = queryNorm
            0.17398985 = fieldWeight in 2347, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2347)
      0.2 = coord(1/5)
    
    Abstract
    Purpose - If an information retrieval system is going to be of value to the user then it must give meaning to the information which matches the meaning given to it by the user. The meaning given to music varies according to who is interpreting it - the author/composer, the performer, cataloguer or the listener - and this affects how music is organized and retrieved. This paper aims to examine the meaning of music, how meaning is communicated and suggests this may affect music retrieval. Design/methodology/approach - Musicology is used to define music and examine its functions leading to a discussion of how music has been organised and described. Various ways of establishing the meaning of music are reviewed, focussing on established musical analysis techniques. It is suggested that traditional methods are of limited use with digitised popular music. A discussion of semiotics and a review of semiotic analysis in western art music leads to a discussion of semiotics of popular music and examines ideas of Middleton, Stefani and Tagg. Findings - Agreeing that music exists when communication takes place, a discussion of selected communication models leads to the proposal of a revised version of Tagg's model, adjusting it to include listener feedback. Originality/value - The outcome of the analysis is a revised version of Tagg's communication model, adapted to reflect user feedback. It is suggested that this revised communication model reflects the way in which meaning is given to music.
  10. MacFarlane, A.; Al-Wabil, A.; Marshall, C.R.; Albrair, A.; Jones, S.A.; Zaphiris, P.: ¬The effect of dyslexia on information retrieval : a pilot study (2010) 0.00
    0.0046599186 = product of:
      0.023299592 = sum of:
        0.023299592 = weight(_text_:system in 3624) [ClassicSimilarity], result of:
          0.023299592 = score(doc=3624,freq=2.0), product of:
            0.13391352 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.04251826 = queryNorm
            0.17398985 = fieldWeight in 3624, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3624)
      0.2 = coord(1/5)
    
    Abstract
    Purpose - The purpose of this paper is to resolve a gap in the knowledge of how people with dyslexia interact with information retrieval (IR) systems, specifically an understanding of their information-searching behaviour. Design/methodology/approach - The dyslexia cognitive profile is used to design a logging system, recording the difference between two sets of participants: dyslexic and control users. A standard Okapi interface is used - together with two standard TREC topics - in order to record the information searching behaviour of these users. Findings - Using the log data, the differences in information-searching behaviour of control and dyslexic users, i.e. in the way the two groups interact with Okapi, are established and it also established that qualitative information collected (such as experience etc.) may not be able to account for these differences. Evidence from query variables was unable to distinguish between groups, but differences on topic for the same variables were recorded. Users who view more documents tended to judge more documents as being relevant, in terms of either the user group or topic. Session data indicated that there may be an important difference between the number of iterations used in a search between the user groups, as there may be little effect from the topic on this variable. Originality/value - This is the first study of the effect of dyslexia on information search behaviour, and it provides some evidence to take the field forward.
  11. MacFarlane, A.; Robertson, S.E.; McCann, J.A.: Parallel computing for passage retrieval (2004) 0.00
    0.0030723398 = product of:
      0.015361699 = sum of:
        0.015361699 = product of:
          0.046085097 = sum of:
            0.046085097 = weight(_text_:22 in 5108) [ClassicSimilarity], result of:
              0.046085097 = score(doc=5108,freq=2.0), product of:
                0.1488917 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04251826 = queryNorm
                0.30952093 = fieldWeight in 5108, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5108)
          0.33333334 = coord(1/3)
      0.2 = coord(1/5)
    
    Date
    20. 1.2007 18:30:22
  12. Inskip, C.; MacFarlane, A.; Rafferty, P.: Organising music for movies (2010) 0.00
    0.001937643 = product of:
      0.009688215 = sum of:
        0.009688215 = product of:
          0.029064644 = sum of:
            0.029064644 = weight(_text_:29 in 3941) [ClassicSimilarity], result of:
              0.029064644 = score(doc=3941,freq=2.0), product of:
                0.14956595 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.04251826 = queryNorm
                0.19432661 = fieldWeight in 3941, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3941)
          0.33333334 = coord(1/3)
      0.2 = coord(1/5)
    
    Date
    29. 8.2010 12:23:57