Search (19 results, page 1 of 1)

  • × author_ss:"Tudhope, D."
  1. Tudhope, D.; Nielsen, M.L.: Introduction to knowledge organization systems and services (2006) 0.07
    0.06672326 = product of:
      0.13344651 = sum of:
        0.06730108 = weight(_text_:services in 5913) [ClassicSimilarity], result of:
          0.06730108 = score(doc=5913,freq=8.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.405633 = fieldWeight in 5913, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5913)
        0.049512394 = weight(_text_:network in 5913) [ClassicSimilarity], result of:
          0.049512394 = score(doc=5913,freq=2.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.2460165 = fieldWeight in 5913, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5913)
        0.016633041 = product of:
          0.033266082 = sum of:
            0.033266082 = weight(_text_:resources in 5913) [ClassicSimilarity], result of:
              0.033266082 = score(doc=5913,freq=2.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.20165458 = fieldWeight in 5913, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5913)
          0.5 = coord(1/2)
      0.5 = coord(3/6)
    
    Abstract
    In a very real sense, this special issue on Knowledge Organization Systems and Services is concerned with new applications, new contexts and new twists of old themes and problems. We are concerned with diverse attempts to apply the outcomes of much work over the years in artificial subject languages and their intellectual structures to facilitate access to digital information in various settings. This issue has its origins in NKOS workshops, held over the last two years in Bath, Vienna, Madrid and Denver, although the majority of contributions resulted from an open call for papers, disseminated in October 2005. NKOS (http://nkos.slis.kent.edu/) is an informal network whose general aim is to enable knowledge organization systems (KOS) to act as networked information services (both machine-to-machine and human facing), supporting the description and retrieval of information resources on the Internet. Since 1997, there has been an NKOS workshop each year, at either the JCDL or ECDL conference (2005 saw an NKOS workshop at both conferences and also at Dublin Core). Previous NKOS-related special issues have appeared in the online Journal of Digital Information in 2001 and 2004 (Hill and Koch 2001, Tudhope and Koch 2004).
    Content
    Einleitung in ein Themenheft "Knowledge organization systems and services"
  2. Tudhope, D.: Knowledge Organization System Services : brief review of NKOS activities and possibility of KOS registries (2007) 0.04
    0.039166186 = product of:
      0.117498554 = sum of:
        0.08076129 = weight(_text_:services in 100) [ClassicSimilarity], result of:
          0.08076129 = score(doc=100,freq=2.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.4867596 = fieldWeight in 100, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.09375 = fieldNorm(doc=100)
        0.036737263 = product of:
          0.07347453 = sum of:
            0.07347453 = weight(_text_:22 in 100) [ClassicSimilarity], result of:
              0.07347453 = score(doc=100,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.46428138 = fieldWeight in 100, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=100)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Date
    22. 9.2007 15:41:14
  3. Binding, C.; Tudhope, D.: Terminology Web services (2010) 0.04
    0.03844414 = product of:
      0.11533242 = sum of:
        0.08076129 = weight(_text_:services in 4067) [ClassicSimilarity], result of:
          0.08076129 = score(doc=4067,freq=8.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.4867596 = fieldWeight in 4067, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.046875 = fieldNorm(doc=4067)
        0.034571126 = product of:
          0.06914225 = sum of:
            0.06914225 = weight(_text_:resources in 4067) [ClassicSimilarity], result of:
              0.06914225 = score(doc=4067,freq=6.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.4191312 = fieldWeight in 4067, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4067)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Controlled terminologies such as classification schemes, name authorities, and thesauri have long been the domain of the library and information science community. Although historically there have been initiatives towards library style classification of web resources, there remain significant problems with searching and quality judgement of online content. Terminology services can play a key role in opening up access to these valuable resources. By exposing controlled terminologies via a web service, organisations maintain data integrity and version control, whilst motivating external users to design innovative ways to present and utilise their data. We introduce terminology web services and review work in the area. We describe the approaches taken in establishing application programming interfaces (API) and discuss the comparative benefits of a dedicated terminology web service versus general purpose programming languages. We discuss experiences at Glamorgan in creating terminology web services and associated client interface components, in particular for the archaeology domain in the STAR (Semantic Technologies for Archaeological Resources) Project.
  4. Tudhope, D.; Hodge, G.: Terminology registries (2007) 0.03
    0.03263849 = product of:
      0.09791546 = sum of:
        0.06730108 = weight(_text_:services in 539) [ClassicSimilarity], result of:
          0.06730108 = score(doc=539,freq=2.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.405633 = fieldWeight in 539, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.078125 = fieldNorm(doc=539)
        0.030614385 = product of:
          0.06122877 = sum of:
            0.06122877 = weight(_text_:22 in 539) [ClassicSimilarity], result of:
              0.06122877 = score(doc=539,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.38690117 = fieldWeight in 539, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=539)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Content
    Präsentation während der Veranstaltung "Networked Knowledge Organization Systems and Services: The 6th European Networked Knowledge Organization Systems (NKOS) Workshop, Workshop at the 11th ECDL Conference, Budapest, Hungary, September 21st 2007".
    Date
    26.12.2011 13:22:07
  5. Golub, K.; Tudhope, D.; Zeng, M.L.; Zumer, M.: Terminology registries for knowledge organization systems : functionality, use, and attributes (2014) 0.03
    0.025158495 = product of:
      0.075475484 = sum of:
        0.057106856 = weight(_text_:services in 1347) [ClassicSimilarity], result of:
          0.057106856 = score(doc=1347,freq=4.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.344191 = fieldWeight in 1347, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.046875 = fieldNorm(doc=1347)
        0.018368632 = product of:
          0.036737263 = sum of:
            0.036737263 = weight(_text_:22 in 1347) [ClassicSimilarity], result of:
              0.036737263 = score(doc=1347,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.23214069 = fieldWeight in 1347, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1347)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Terminology registries (TRs) are a crucial element of the infrastructure required for resource discovery services, digital libraries, Linked Data, and semantic interoperability generally. They can make the content of knowledge organization systems (KOS) available both for human and machine access. The paper describes the attributes and functionality for a TR, based on a review of published literature, existing TRs, and a survey of experts. A domain model based on user tasks is constructed and a set of core metadata elements for use in TRs is proposed. Ideally, the TR should allow searching as well as browsing for a KOS, matching a user's search while also providing information about existing terminology services, accessible to both humans and machines. The issues surrounding metadata for KOS are also discussed, together with the rationale for different aspects and the importance of a core set of KOS metadata for future machine-based access; a possible core set of metadata elements is proposed. This is dealt with in terms of practical experience and in relation to the Dublin Core Application Profile.
    Date
    22. 8.2014 17:12:54
  6. Tudhope, D.; Binding, C.: Toward terminology services : experiences with a pilot Web service thesaurus browser (2006) 0.02
    0.019977996 = product of:
      0.059933987 = sum of:
        0.046627555 = weight(_text_:services in 1955) [ClassicSimilarity], result of:
          0.046627555 = score(doc=1955,freq=6.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.2810308 = fieldWeight in 1955, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.03125 = fieldNorm(doc=1955)
        0.013306432 = product of:
          0.026612865 = sum of:
            0.026612865 = weight(_text_:resources in 1955) [ClassicSimilarity], result of:
              0.026612865 = score(doc=1955,freq=2.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.16132367 = fieldWeight in 1955, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1955)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Dublin Core recommends controlled terminology for the subject of a resource. Knowledge organization systems (KOS), such as classifications, gazetteers, taxonomies and thesauri, provide controlled vocabularies that organize and structure concepts for indexing, classifying, browsing and search. For example, a thesaurus employs a set of standard semantic relationships (ISO 2788, ISO 5964), and major thesauri have a large entry vocabulary of terms considered equivalent for retrieval purposes. Many KOS have been made available for Web-based access. However, they are often not fully integrated into indexing and search systems and the full potential for networked and programmatic access remains untapped. The lack of standardized access and interchange formats impedes wider use of KOS resources. We developed a Web demonstrator (www.comp.glam.ac.uk/~FACET/webdemo/) for the FACET project (www.comp.glam.ac.uk/~facet/facetproject.html) that explored thesaurus-based query expansion with the Getty Art and Architecture Thesaurus. A Web demonstrator was implemented via Active Server Pages (ASP) with server-side scripting and compiled server-side components for database access, and cascading style sheets for presentation. The browser-based interactive interface permits dynamic control of query term expansion. However, being based on a custom thesaurus representation and API, the techniques cannot be applied directly to thesauri in other formats on the Web. General programmatic access requires commonly agreed protocols, for example, building on Web and Grid services. The development of common KOS representation formats and service protocols are closely linked. Linda Hill and colleagues argued in 2002 for a general KOS service protocol from which protocols for specific types of KOS can be derived. Thus, in the future, a combination of thesaurus and query protocols might permit a thesaurus to be used with a choice of search tools on various kinds of databases. Service-oriented architectures bring an opportunity for moving toward a clearer separation of interface components from the underlying data sources. In our view, basing distributed protocol services on the atomic elements of thesaurus data structures and relationships is not necessarily the best approach because client operations that require multiple client-server calls would carry too much overhead. This would limit the interfaces that could be offered by applications following such a protocol. Advanced interactive interfaces require protocols that group primitive thesaurus data elements (via their relationships) into composites to achieve reasonable response.
  7. Tudhope, D.; Binding, C.: Mapping between linked data vocabularies in ARIADNE (2015) 0.02
    0.01905774 = product of:
      0.057173215 = sum of:
        0.03365054 = weight(_text_:services in 2250) [ClassicSimilarity], result of:
          0.03365054 = score(doc=2250,freq=2.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.2028165 = fieldWeight in 2250, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2250)
        0.023522673 = product of:
          0.047045346 = sum of:
            0.047045346 = weight(_text_:resources in 2250) [ClassicSimilarity], result of:
              0.047045346 = score(doc=2250,freq=4.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.28518265 = fieldWeight in 2250, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2250)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Semantic Enrichment Enabling Sustainability of Archaeological Links (SENESCHAL) was a project coordinated by the Hypermedia Research Unit at the University of South Wales. The project aims included widening access to key vocabulary resources. National cultural heritage thesauri and vocabularies are used by both national organizations and local authority Historic Environment Records and could potentially act as vocabulary hubs for the Web of Data. Following completion, a set of prominent UK archaeological thesauri and vocabularies is now freely available as Linked Open Data (LOD) via http://www.heritagedata.org - together with open source web services and user interface controls. This presentation will reflect on work done to date for the ARIADNE FP7 infrastructure project (http://www.ariadne-infrastructure.eu) mapping between archaeological vocabularies in different languages and the utility of a hub architecture. The poly-hierarchical structure of the Getty Art & Architecture Thesaurus (AAT) was extracted for use as an example mediating structure to interconnect various multilingual vocabularies originating from ARIADNE data providers. Vocabulary resources were first converted to a common concept-based format (SKOS) and the concepts were then manually mapped to nodes of the extracted AAT structure using some judgement on the meaning of terms and scope notes. Results are presented along with reflections on the wider application to existing European archaeological vocabularies and associated online datasets.
  8. Golub, K.; Soergel, D.; Buchanan, G.; Tudhope, D.; Lykke, M.; Hiom, D.: ¬A framework for evaluating automatic indexing or classification in the context of retrieval (2016) 0.02
    0.0166584 = product of:
      0.049975198 = sum of:
        0.033342157 = weight(_text_:computer in 3311) [ClassicSimilarity], result of:
          0.033342157 = score(doc=3311,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.20188503 = fieldWeight in 3311, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3311)
        0.016633041 = product of:
          0.033266082 = sum of:
            0.033266082 = weight(_text_:resources in 3311) [ClassicSimilarity], result of:
              0.033266082 = score(doc=3311,freq=2.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.20165458 = fieldWeight in 3311, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3311)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Tools for automatic subject assignment help deal with scale and sustainability in creating and enriching metadata, establishing more connections across and between resources and enhancing consistency. Although some software vendors and experimental researchers claim the tools can replace manual subject indexing, hard scientific evidence of their performance in operating information environments is scarce. A major reason for this is that research is usually conducted in laboratory conditions, excluding the complexities of real-life systems and situations. The article reviews and discusses issues with existing evaluation approaches such as problems of aboutness and relevance assessments, implying the need to use more than a single "gold standard" method when evaluating indexing and retrieval, and proposes a comprehensive evaluation framework. The framework is informed by a systematic review of the literature on evaluation approaches: evaluating indexing quality directly through assessment by an evaluator or through comparison with a gold standard, evaluating the quality of computer-assisted indexing directly in the context of an indexing workflow, and evaluating indexing quality indirectly through analyzing retrieval performance.
  9. Tudhope, D.; Blocks, D.; Cunliffe, D.; Binding, C.: Query expansion via conceptual distance in thesaurus indexed collections (2006) 0.01
    0.008252066 = product of:
      0.049512394 = sum of:
        0.049512394 = weight(_text_:network in 2215) [ClassicSimilarity], result of:
          0.049512394 = score(doc=2215,freq=2.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.2460165 = fieldWeight in 2215, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2215)
      0.16666667 = coord(1/6)
    
    Abstract
    Purpose - The purpose of this paper is to explore query expansion via conceptual distance in thesaurus indexed collections Design/methodology/approach - An extract of the National Museum of Science and Industry's collections database, indexed with the Getty Art and Architecture Thesaurus (AAT), was the dataset for the research. The system architecture and algorithms for semantic closeness and the matching function are outlined. Standalone and web interfaces are described and formative qualitative user studies are discussed. One user session is discussed in detail, together with a scenario based on a related public inquiry. Findings are set in context of the literature on thesaurus-based query expansion. This paper discusses the potential of query expansion techniques using the semantic relationships in a faceted thesaurus. Findings - Thesaurus-assisted retrieval systems have potential for multi-concept descriptors, permitting very precise queries and indexing. However, indexer and searcher may differ in terminology judgments and there may not be any exactly matching results. The integration of semantic closeness in the matching function permits ranked results for multi-concept queries in thesaurus-indexed applications. An in-memory representation of the thesaurus semantic network allows a combination of automatic and interactive control of expansion and control of expansion on individual query terms. Originality/value - The application of semantic expansion to browsing may be useful in interface options where thesaurus structure is hidden.
  10. Tudhope, D.; Binding, C.: Still quite popular after all those years : the continued relevance of the information retrieval thesaurus (2016) 0.01
    0.0067301076 = product of:
      0.040380646 = sum of:
        0.040380646 = weight(_text_:services in 2908) [ClassicSimilarity], result of:
          0.040380646 = score(doc=2908,freq=2.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.2433798 = fieldWeight in 2908, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.046875 = fieldNorm(doc=2908)
      0.16666667 = coord(1/6)
    
    Abstract
    The recent ISKO-UK conference considered the question of whether the traditional thesaurus has any place in modern information retrieval. This note is intended to continue in the spirit of that good-natured debate, arguing that there is indeed a role today and highlighting some recent work showing the continued relevance of the thesaurus, particularly in the linked data area. Key functionality that a thesaurus makes possible is discussed. A brief outline is provided of prominent work hat employs thesauri in three key areas of infrastructure underpinning advanced retrieval functionality today: metadata enrichment,vocabulary mapping and web services.
  11. Tudhope, D.; Binding, C.; Blocks, D.; Cunliffe, D.: FACET: thesaurus retrieval with semantic term expansion (2002) 0.01
    0.006601652 = product of:
      0.039609913 = sum of:
        0.039609913 = weight(_text_:network in 175) [ClassicSimilarity], result of:
          0.039609913 = score(doc=175,freq=2.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.1968132 = fieldWeight in 175, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.03125 = fieldNorm(doc=175)
      0.16666667 = coord(1/6)
    
    Abstract
    There are many advantages for Digital Libraries in indexing with classifications or thesauri, but some current disincentive in the lack of flexible retrieval tools that deal with compound descriptors. This demonstration of a research prototype illustrates a matching function for compound descriptors, or multi-concept subject headings, that does not rely on exact matching but incorporates term expansion via thesaurus semantic relationships to produce ranked results that take account of missing and partially matching terms. The matching function is based on a measure of semantic closeness between terms.The work is part of the EPSRC funded FACET project in collaboration with the UK National Museum of Science and Industry (NMSI) which includes the National Railway Museum. An export of NMSI's Collections Database is used as the dataset for the research. The J. Paul Getty Trust's Art and Architecture Thesaurus (AAT) is the main thesaurus in the project. The AAT is a widely used thesaurus (over 120,000 terms). Descriptors are organised in 7 facets representing separate conceptual classes of terms.The FACET application is a multi tiered architecture accessing a SQL Server database, with an OLE DB connection. The thesauri are stored as relational tables in the Server's database. However, a key component of the system is a parallel representation of the underlying semantic network as an in-memory structure of thesaurus concepts (corresponding to preferred terms). The structure models the hierarchical and associative interrelationships of thesaurus concepts via weighted poly-hierarchical links. Its primary purpose is real-time semantic expansion of query terms, achieved by a spreading activation semantic closeness algorithm. Queries with associated results are stored persistently using XML format data. A Visual Basic interface combines a thesaurus browser and an initial term search facility that takes into account equivalence relationships. Terms are dragged to a direct manipulation Query Builder which maintains the facet structure.
  12. Vlachidis, A.; Binding, C.; Tudhope, D.; May, K.: Excavating grey literature : a case study on the rich indexing of archaeological documents via natural language-processing techniques and knowledge-based resources (2010) 0.01
    0.0058675855 = product of:
      0.035205513 = sum of:
        0.035205513 = product of:
          0.07041103 = sum of:
            0.07041103 = weight(_text_:resources in 3948) [ClassicSimilarity], result of:
              0.07041103 = score(doc=3948,freq=14.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.42682233 = fieldWeight in 3948, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3948)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    Purpose - This paper sets out to discuss the use of information extraction (IE), a natural language-processing (NLP) technique to assist "rich" semantic indexing of diverse archaeological text resources. The focus of the research is to direct a semantic-aware "rich" indexing of diverse natural language resources with properties capable of satisfying information retrieval from online publications and datasets associated with the Semantic Technologies for Archaeological Resources (STAR) project. Design/methodology/approach - The paper proposes use of the English Heritage extension (CRM-EH) of the standard core ontology in cultural heritage, CIDOC CRM, and exploitation of domain thesauri resources for driving and enhancing an Ontology-Oriented Information Extraction process. The process of semantic indexing is based on a rule-based Information Extraction technique, which is facilitated by the General Architecture of Text Engineering (GATE) toolkit and expressed by Java Annotation Pattern Engine (JAPE) rules. Findings - Initial results suggest that the combination of information extraction with knowledge resources and standard conceptual models is capable of supporting semantic-aware term indexing. Additional efforts are required for further exploitation of the technique and adoption of formal evaluation methods for assessing the performance of the method in measurable terms. Originality/value - The value of the paper lies in the semantic indexing of 535 unpublished online documents often referred to as "Grey Literature", from the Archaeological Data Service OASIS corpus (Online AccesS to the Index of archaeological investigationS), with respect to the CRM ontological concepts E49.Time Appellation and P19.Physical Object.
    Footnote
    Beitrag in einem Special Issue: Content architecture: exploiting and managing diverse resources: proceedings of the first national conference of the United Kingdom chapter of the International Society for Knowedge Organization (ISKO)
  13. Binding, C.; Tudhope, D.: Improving interoperability using vocabulary linked data (2015) 0.00
    0.0048015458 = product of:
      0.028809274 = sum of:
        0.028809274 = product of:
          0.057618547 = sum of:
            0.057618547 = weight(_text_:resources in 2205) [ClassicSimilarity], result of:
              0.057618547 = score(doc=2205,freq=6.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.349276 = fieldWeight in 2205, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2205)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    The concept of Linked Data has been an emerging theme within the computing and digital heritage areas in recent years. The growth and scale of Linked Data has underlined the need for greater commonality in concept referencing, to avoid local redefinition and duplication of reference resources. Achieving domain-wide agreement on common vocabularies would be an unreasonable expectation; however, datasets often already have local vocabulary resources defined, and so the prospects for large-scale interoperability can be substantially improved by creating alignment links from these local vocabularies out to common external reference resources. The ARIADNE project is undertaking large-scale integration of archaeology dataset metadata records, to create a cross-searchable research repository resource. Key to enabling this cross search will be the 'subject' metadata originating from multiple data providers, containing terms from multiple multilingual controlled vocabularies. This paper discusses various aspects of vocabulary mapping. Experience from the previous SENESCHAL project in the publication of controlled vocabularies as Linked Open Data is discussed, emphasizing the importance of unique URI identifiers for vocabulary concepts. There is a need to align legacy indexing data to the uniquely defined concepts and examples are discussed of SENESCHAL data alignment work. A case study for the ARIADNE project presents work on mapping between vocabularies, based on the Getty Art and Architecture Thesaurus as a central hub and employing an interactive vocabulary mapping tool developed for the project, which generates SKOS mapping relationships in JSON and other formats. The potential use of such vocabulary mappings to assist cross search over archaeological datasets from different countries is illustrated in a pilot experiment. The results demonstrate the enhanced opportunities for interoperability and cross searching that the approach offers.
  14. Matthews, B.; Jones, C.; Puzon, B.; Moon, J.; Tudhope, D.; Golub, K.; Nielsen, M.L.: ¬An evaluation of enhancing social tagging with a knowledge organization system (2010) 0.00
    0.0039204457 = product of:
      0.023522673 = sum of:
        0.023522673 = product of:
          0.047045346 = sum of:
            0.047045346 = weight(_text_:resources in 4171) [ClassicSimilarity], result of:
              0.047045346 = score(doc=4171,freq=4.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.28518265 = fieldWeight in 4171, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4171)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    Purpose - Traditional subject indexing and classification are considered infeasible in many digital collections. This paper seeks to investigate ways of enhancing social tagging via knowledge organization systems, with a view to improving the quality of tags for increased information discovery and retrieval performance. Design/methodology/approach - Enhanced tagging interfaces were developed for exemplar online repositories, and trials were undertaken with author and reader groups to evaluate the effectiveness of tagging augmented with control vocabulary for subject indexing of papers in online repositories. Findings - The results showed that using a knowledge organisation system to augment tagging does appear to increase the effectiveness of non-specialist users (that is, without information science training) in subject indexing. Research limitations/implications - While limited by the size and scope of the trials undertaken, these results do point to the usefulness of a mixed approach in supporting the subject indexing of online resources. Originality/value - The value of this work is as a guide to future developments in the practical support for resource indexing in online repositories.
    Footnote
    Beitrag in einem Special Issue: Content architecture: exploiting and managing diverse resources: proceedings of the first national conference of the United Kingdom chapter of the International Society for Knowedge Organization (ISKO)
  15. Vlachidis, A.; Tudhope, D.: ¬A knowledge-based approach to information extraction for semantic interoperability in the archaeology domain (2016) 0.00
    0.0039204457 = product of:
      0.023522673 = sum of:
        0.023522673 = product of:
          0.047045346 = sum of:
            0.047045346 = weight(_text_:resources in 2895) [ClassicSimilarity], result of:
              0.047045346 = score(doc=2895,freq=4.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.28518265 = fieldWeight in 2895, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2895)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    The article presents a method for automatic semantic indexing of archaeological grey-literature reports using empirical (rule-based) Information Extraction techniques in combination with domain-specific knowledge organization systems. The semantic annotation system (OPTIMA) performs the tasks of Named Entity Recognition, Relation Extraction, Negation Detection, and Word-Sense Disambiguation using hand-crafted rules and terminological resources for associating contextual abstractions with classes of the standard ontology CIDOC Conceptual Reference Model (CRM) for cultural heritage and its archaeological extension, CRM-EH. Relation Extraction (RE) performance benefits from a syntactic-based definition of RE patterns derived from domain oriented corpus analysis. The evaluation also shows clear benefit in the use of assistive natural language processing (NLP) modules relating to Word-Sense Disambiguation, Negation Detection, and Noun Phrase Validation, together with controlled thesaurus expansion. The semantic indexing results demonstrate the capacity of rule-based Information Extraction techniques to deliver interoperable semantic abstractions (semantic annotations) with respect to the CIDOC CRM and archaeological thesauri. Major contributions include recognition of relevant entities using shallow parsing NLP techniques driven by a complimentary use of ontological and terminological domain resources and empirical derivation of context-driven RE rules for the recognition of semantic relationships from phrases of unstructured text.
  16. Golub, K.; Lykke, M.; Tudhope, D.: Enhancing social tagging with automated keywords from the Dewey Decimal Classification (2014) 0.00
    0.0027721736 = product of:
      0.016633041 = sum of:
        0.016633041 = product of:
          0.033266082 = sum of:
            0.033266082 = weight(_text_:resources in 2918) [ClassicSimilarity], result of:
              0.033266082 = score(doc=2918,freq=2.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.20165458 = fieldWeight in 2918, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2918)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    Purpose - The purpose of this paper is to explore the potential of applying the Dewey Decimal Classification (DDC) as an established knowledge organization system (KOS) for enhancing social tagging, with the ultimate purpose of improving subject indexing and information retrieval. Design/methodology/approach - Over 11.000 Intute metadata records in politics were used. Totally, 28 politics students were each given four tasks, in which a total of 60 resources were tagged in two different configurations, one with uncontrolled social tags only and another with uncontrolled social tags as well as suggestions from a controlled vocabulary. The controlled vocabulary was DDC comprising also mappings from the Library of Congress Subject Headings. Findings - The results demonstrate the importance of controlled vocabulary suggestions for indexing and retrieval: to help produce ideas of which tags to use, to make it easier to find focus for the tagging, to ensure consistency and to increase the number of access points in retrieval. The value and usefulness of the suggestions proved to be dependent on the quality of the suggestions, both as to conceptual relevance to the user and as to appropriateness of the terminology. Originality/value - No research has investigated the enhancement of social tagging with suggestions from the DDC, an established KOS, in a user trial, comparing social tagging only and social tagging enhanced with the suggestions. This paper is a final reflection on all aspects of the study.
  17. Golub, K.; Hansson, J.; Soergel, D.; Tudhope, D.: Managing classification in libraries : a methodological outline for evaluating automatic subject indexing and classification in Swedish library catalogues (2015) 0.00
    0.0027721736 = product of:
      0.016633041 = sum of:
        0.016633041 = product of:
          0.033266082 = sum of:
            0.033266082 = weight(_text_:resources in 2300) [ClassicSimilarity], result of:
              0.033266082 = score(doc=2300,freq=2.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.20165458 = fieldWeight in 2300, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2300)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    Subject terms play a crucial role in resource discovery but require substantial effort to produce. Automatic subject classification and indexing address problems of scale and sustainability and can be used to enrich existing bibliographic records, establish more connections across and between resources and enhance consistency of bibliographic data. The paper aims to put forward a complex methodological framework to evaluate automatic classification tools of Swedish textual documents based on the Dewey Decimal Classification (DDC) recently introduced to Swedish libraries. Three major complementary approaches are suggested: a quality-built gold standard, retrieval effects, domain analysis. The gold standard is built based on input from at least two catalogue librarians, end-users expert in the subject, end users inexperienced in the subject and automated tools. Retrieval effects are studied through a combination of assigned and free tasks, including factual and comprehensive types. The study also takes into consideration the different role and character of subject terms in various knowledge domains, such as scientific disciplines. As a theoretical framework, domain analysis is used and applied in relation to the implementation of DDC in Swedish libraries and chosen domains of knowledge within the DDC itself.
  18. Khoo, M.J.; Ahn, J.-w.; Binding, C.; Jones, H.J.; Lin, X.; Massam, D.; Tudhope, D.: Augmenting Dublin Core digital library metadata with Dewey Decimal Classification (2015) 0.00
    0.002217739 = product of:
      0.013306432 = sum of:
        0.013306432 = product of:
          0.026612865 = sum of:
            0.026612865 = weight(_text_:resources in 2320) [ClassicSimilarity], result of:
              0.026612865 = score(doc=2320,freq=2.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.16132367 = fieldWeight in 2320, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2320)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    Purpose - The purpose of this paper is to describe a new approach to a well-known problem for digital libraries, how to search across multiple unrelated libraries with a single query. Design/methodology/approach - The approach involves creating new Dewey Decimal Classification terms and numbers from existing Dublin Core records. In total, 263,550 records were harvested from three digital libraries. Weighted key terms were extracted from the title, description and subject fields of each record. Ranked DDC classes were automatically generated from these key terms by considering DDC hierarchies via a series of filtering and aggregation stages. A mean reciprocal ranking evaluation compared a sample of 49 generated classes against DDC classes created by a trained librarian for the same records. Findings - The best results combined weighted key terms from the title, description and subject fields. Performance declines with increased specificity of DDC level. The results compare favorably with similar studies. Research limitations/implications - The metadata harvest required manual intervention and the evaluation was resource intensive. Future research will look at evaluation methodologies that take account of issues of consistency and ecological validity. Practical implications - The method does not require training data and is easily scalable. The pipeline can be customized for individual use cases, for example, recall or precision enhancing. Social implications - The approach can provide centralized access to information from multiple domains currently provided by individual digital libraries. Originality/value - The approach addresses metadata normalization in the context of web resources. The automatic classification approach accounts for matches within hierarchies, aggregating lower level matches to broader parents and thus approximates the practices of a human cataloger.
  19. Jones, I.; Cunliffe, D.; Tudhope, D.: Natural language processing and knowledge organization systems as an aid to retrieval (2004) 0.00
    0.0019405214 = product of:
      0.0116431285 = sum of:
        0.0116431285 = product of:
          0.023286257 = sum of:
            0.023286257 = weight(_text_:resources in 2677) [ClassicSimilarity], result of:
              0.023286257 = score(doc=2677,freq=2.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.14115821 = fieldWeight in 2677, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=2677)
          0.5 = coord(1/2)
      0.16666667 = coord(1/6)
    
    Abstract
    This paper discusses research that employs methods from Natural Language Processing (NLP) in exploiting the intellectual resources of Knowledge Organization Systems (KOS), particularly in the retrieval of information. A technique for the disambiguation of homographs and nominal compounds in free text, where these are known ambiguous terms in the KOS itself, is described. The use of Roget's Thesaurus as an intermediary in the process is also reported. A short review of the relevant literature in the field is given. Design considerations, results and conclusions are presented from the implementation of a prototype system. The linguistic techniques are applied at two complementary levels, namely an a free text string used as an entry point to the KOS, and an the underlying controlled vocabulary itself.