Search (27 results, page 2 of 2)

  • × author_ss:"Tudhope, D."
  • × type_ss:"a"
  1. Tudhope, D.; Taylor, C.: ¬A unified similarity coefficient for navigating through multi-dimensional information (1996) 0.00
    2.497293E-4 = product of:
      0.0049945856 = sum of:
        0.0049945856 = weight(_text_:in in 7460) [ClassicSimilarity], result of:
          0.0049945856 = score(doc=7460,freq=4.0), product of:
            0.039165888 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.02879306 = queryNorm
            0.12752387 = fieldWeight in 7460, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=7460)
      0.05 = coord(1/20)
    
    Abstract
    Describes an integrated approach to similarity coefficients for information spaces with multiple dimensions of different types of index term. Categorises applications of similarity coefficients underlying different navigation tools in hypermedia by type of term. Describes an implementation of a unified similarity coefficient based on work in numerical taxonomy, with illustrative scenarios from an experimental navigation via similarity tool for a prototype social history museum hypermedia system. The underlying architecture is based on a semantic approach, where semantic relationships can exist between index terms. This allows imprecise matching when comparing for similarity, with distance measures yielding a degree of match. A ranked list of matching items over several weighted dimensions is returned by the similarity navigation tool. The approach has the potential of allowing different access methods to multimedia data to be combined
  2. Binding, C.; Tudhope, D.: Integrating faceted structure into the search process (2004) 0.00
    2.497293E-4 = product of:
      0.0049945856 = sum of:
        0.0049945856 = weight(_text_:in in 2627) [ClassicSimilarity], result of:
          0.0049945856 = score(doc=2627,freq=4.0), product of:
            0.039165888 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.02879306 = queryNorm
            0.12752387 = fieldWeight in 2627, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=2627)
      0.05 = coord(1/20)
    
    Abstract
    The nature of search requirements is perceived to be changing, fuelled by a growing dissatisfaction with the marginal accuracy and often overwhelming quantity of results from simple keyword matching techniques. Traditional search interfaces fail to acknowledge and utilise the implicit underlying structure present within a typical keyword query. Faceted structure can (and should) perform a significant role in this area - acting as the basis for mediation between searcher and indexer, and guiding query formulation and reformulation by interactively educating the user about the native domain. This paper discusses the possible benefits of applying faceted knowledge organization systems to enhance query structure, query visualisation and the overall query process, drawing an the outcomes of a recently completed research project.
    Series
    Advances in knowledge organization; vol.9
  3. Tudhope, D.; Binding, C.: Toward terminology services : experiences with a pilot Web service thesaurus browser (2006) 0.00
    2.35447E-4 = product of:
      0.00470894 = sum of:
        0.00470894 = weight(_text_:in in 1955) [ClassicSimilarity], result of:
          0.00470894 = score(doc=1955,freq=8.0), product of:
            0.039165888 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.02879306 = queryNorm
            0.120230645 = fieldWeight in 1955, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.03125 = fieldNorm(doc=1955)
      0.05 = coord(1/20)
    
    Abstract
    Dublin Core recommends controlled terminology for the subject of a resource. Knowledge organization systems (KOS), such as classifications, gazetteers, taxonomies and thesauri, provide controlled vocabularies that organize and structure concepts for indexing, classifying, browsing and search. For example, a thesaurus employs a set of standard semantic relationships (ISO 2788, ISO 5964), and major thesauri have a large entry vocabulary of terms considered equivalent for retrieval purposes. Many KOS have been made available for Web-based access. However, they are often not fully integrated into indexing and search systems and the full potential for networked and programmatic access remains untapped. The lack of standardized access and interchange formats impedes wider use of KOS resources. We developed a Web demonstrator (www.comp.glam.ac.uk/~FACET/webdemo/) for the FACET project (www.comp.glam.ac.uk/~facet/facetproject.html) that explored thesaurus-based query expansion with the Getty Art and Architecture Thesaurus. A Web demonstrator was implemented via Active Server Pages (ASP) with server-side scripting and compiled server-side components for database access, and cascading style sheets for presentation. The browser-based interactive interface permits dynamic control of query term expansion. However, being based on a custom thesaurus representation and API, the techniques cannot be applied directly to thesauri in other formats on the Web. General programmatic access requires commonly agreed protocols, for example, building on Web and Grid services. The development of common KOS representation formats and service protocols are closely linked. Linda Hill and colleagues argued in 2002 for a general KOS service protocol from which protocols for specific types of KOS can be derived. Thus, in the future, a combination of thesaurus and query protocols might permit a thesaurus to be used with a choice of search tools on various kinds of databases. Service-oriented architectures bring an opportunity for moving toward a clearer separation of interface components from the underlying data sources. In our view, basing distributed protocol services on the atomic elements of thesaurus data structures and relationships is not necessarily the best approach because client operations that require multiple client-server calls would carry too much overhead. This would limit the interfaces that could be offered by applications following such a protocol. Advanced interactive interfaces require protocols that group primitive thesaurus data elements (via their relationships) into composites to achieve reasonable response.
  4. Vlachidis, A.; Binding, C.; Tudhope, D.; May, K.: Excavating grey literature : a case study on the rich indexing of archaeological documents via natural language-processing techniques and knowledge-based resources (2010) 0.00
    2.35447E-4 = product of:
      0.00470894 = sum of:
        0.00470894 = weight(_text_:in in 3948) [ClassicSimilarity], result of:
          0.00470894 = score(doc=3948,freq=8.0), product of:
            0.039165888 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.02879306 = queryNorm
            0.120230645 = fieldWeight in 3948, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.03125 = fieldNorm(doc=3948)
      0.05 = coord(1/20)
    
    Abstract
    Purpose - This paper sets out to discuss the use of information extraction (IE), a natural language-processing (NLP) technique to assist "rich" semantic indexing of diverse archaeological text resources. The focus of the research is to direct a semantic-aware "rich" indexing of diverse natural language resources with properties capable of satisfying information retrieval from online publications and datasets associated with the Semantic Technologies for Archaeological Resources (STAR) project. Design/methodology/approach - The paper proposes use of the English Heritage extension (CRM-EH) of the standard core ontology in cultural heritage, CIDOC CRM, and exploitation of domain thesauri resources for driving and enhancing an Ontology-Oriented Information Extraction process. The process of semantic indexing is based on a rule-based Information Extraction technique, which is facilitated by the General Architecture of Text Engineering (GATE) toolkit and expressed by Java Annotation Pattern Engine (JAPE) rules. Findings - Initial results suggest that the combination of information extraction with knowledge resources and standard conceptual models is capable of supporting semantic-aware term indexing. Additional efforts are required for further exploitation of the technique and adoption of formal evaluation methods for assessing the performance of the method in measurable terms. Originality/value - The value of the paper lies in the semantic indexing of 535 unpublished online documents often referred to as "Grey Literature", from the Archaeological Data Service OASIS corpus (Online AccesS to the Index of archaeological investigationS), with respect to the CRM ontological concepts E49.Time Appellation and P19.Physical Object.
    Footnote
    Beitrag in einem Special Issue: Content architecture: exploiting and managing diverse resources: proceedings of the first national conference of the United Kingdom chapter of the International Society for Knowedge Organization (ISKO)
  5. Souza, R.R.; Tudhope, D.; Almeida, M.B.: Towards a taxonomy of KOS (2012) 0.00
    2.0601612E-4 = product of:
      0.0041203224 = sum of:
        0.0041203224 = weight(_text_:in in 139) [ClassicSimilarity], result of:
          0.0041203224 = score(doc=139,freq=2.0), product of:
            0.039165888 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.02879306 = queryNorm
            0.10520181 = fieldWeight in 139, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.0546875 = fieldNorm(doc=139)
      0.05 = coord(1/20)
    
    Abstract
    This paper analyzes previous work on the classification of Knowledge Organization Systems (KOS), discusses strengths and weaknesses, and proposes a new and integrative framework. It argues that current analyses of the KOS tend to be idiosyncratic and incomplete, relying on a limited number of dimensions of analysis. The paper discusses why and how KOS should be classified on a new basis. Based on the available literature and previous work, the authors propose a wider set of dimensions for the analysis of KOS. These are represented in a taxonomy of KOS. Issues arising are discussed.
  6. Tudhope, D.; Taylor, C.: Navigation via similarity (1997) 0.00
    1.7658525E-4 = product of:
      0.003531705 = sum of:
        0.003531705 = weight(_text_:in in 155) [ClassicSimilarity], result of:
          0.003531705 = score(doc=155,freq=2.0), product of:
            0.039165888 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.02879306 = queryNorm
            0.09017298 = fieldWeight in 155, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.046875 = fieldNorm(doc=155)
      0.05 = coord(1/20)
    
    Abstract
    Describes a research project, in which similarity measures have been extended to include imprecise matching over different dimensions of structured classification schemes (subject, soace, time). The semantic similarity of information units forms the basis for the automatic construction of links and is integrated into hypermedia navigation. Outlines a semantic hypermedia architecture, and a prototype museum social history application. Presents illustrative navigation scenarios which make use of a navigation via similarity tool. The temporal mesaures of semantic closeness underpin the similarity tool. The temporal measures takes account of periods as well as time points. The most general measure is based on a traversal of a semantic net, taking into account relationship type and level of specialisation. It is based on a notion of closeness rather than absolute distance, and returns a seit of semantically close terms. Discusses a methods of calculating semantic similarity between sets of index terms, based on the maximal closeness values achieved by each term
  7. Khoo, M.J.; Ahn, J.-w.; Binding, C.; Jones, H.J.; Lin, X.; Massam, D.; Tudhope, D.: Augmenting Dublin Core digital library metadata with Dewey Decimal Classification (2015) 0.00
    1.6648618E-4 = product of:
      0.0033297236 = sum of:
        0.0033297236 = weight(_text_:in in 2320) [ClassicSimilarity], result of:
          0.0033297236 = score(doc=2320,freq=4.0), product of:
            0.039165888 = queryWeight, product of:
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.02879306 = queryNorm
            0.08501591 = fieldWeight in 2320, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.3602545 = idf(docFreq=30841, maxDocs=44218)
              0.03125 = fieldNorm(doc=2320)
      0.05 = coord(1/20)
    
    Abstract
    Purpose - The purpose of this paper is to describe a new approach to a well-known problem for digital libraries, how to search across multiple unrelated libraries with a single query. Design/methodology/approach - The approach involves creating new Dewey Decimal Classification terms and numbers from existing Dublin Core records. In total, 263,550 records were harvested from three digital libraries. Weighted key terms were extracted from the title, description and subject fields of each record. Ranked DDC classes were automatically generated from these key terms by considering DDC hierarchies via a series of filtering and aggregation stages. A mean reciprocal ranking evaluation compared a sample of 49 generated classes against DDC classes created by a trained librarian for the same records. Findings - The best results combined weighted key terms from the title, description and subject fields. Performance declines with increased specificity of DDC level. The results compare favorably with similar studies. Research limitations/implications - The metadata harvest required manual intervention and the evaluation was resource intensive. Future research will look at evaluation methodologies that take account of issues of consistency and ecological validity. Practical implications - The method does not require training data and is easily scalable. The pipeline can be customized for individual use cases, for example, recall or precision enhancing. Social implications - The approach can provide centralized access to information from multiple domains currently provided by individual digital libraries. Originality/value - The approach addresses metadata normalization in the context of web resources. The automatic classification approach accounts for matches within hierarchies, aggregating lower level matches to broader parents and thus approximates the practices of a human cataloger.