Search (151 results, page 1 of 8)

  • × type_ss:"a"
  • × type_ss:"el"
  • × year_i:[2010 TO 2020}
  1. Klic, L.; Miller, M.; Nelson, J.K.; Pattuelli, C.; Provo, A.: ¬The drawings of the Florentine painters : from print catalog to linked open data (2017) 0.03
    0.032600008 = product of:
      0.14670004 = sum of:
        0.0921318 = weight(_text_:readable in 4105) [ClassicSimilarity], result of:
          0.0921318 = score(doc=4105,freq=2.0), product of:
            0.2262076 = queryWeight, product of:
              6.1439276 = idf(docFreq=257, maxDocs=44218)
              0.036818076 = queryNorm
            0.4072887 = fieldWeight in 4105, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.1439276 = idf(docFreq=257, maxDocs=44218)
              0.046875 = fieldNorm(doc=4105)
        0.054568242 = weight(_text_:data in 4105) [ClassicSimilarity], result of:
          0.054568242 = score(doc=4105,freq=10.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.46871632 = fieldWeight in 4105, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046875 = fieldNorm(doc=4105)
      0.22222222 = coord(2/9)
    
    Abstract
    The Drawings of The Florentine Painters project created the first online database of Florentine Renaissance drawings by applying Linked Open Data (LOD) techniques to a foundational text of the same name, first published by Bernard Berenson in 1903 (revised and expanded editions, 1938 and 1961). The goal was to make Berenson's catalog information-still an essential information resource today-available in a machine-readable format, allowing researchers to access the source content through open data services. This paper provides a technical overview of the methods and processes applied in the conversion of Berenson's catalog to LOD using the CIDOC-CRM ontology; it also discusses the different phases of the project, focusing on the challenges and issues of data transformation and publishing. The project was funded by the Samuel H. Kress Foundation and organized by Villa I Tatti, The Harvard University Center for Italian Renaissance Studies. Catalog: http://florentinedrawings.itatti.harvard.edu. Data Endpoint: http://data.itatti.harvard.edu.
  2. Bandholtz, T.; Schulte-Coerne, T.; Glaser, R.; Fock, J.; Keller, T.: iQvoc - open source SKOS(XL) maintenance and publishing tool (2010) 0.03
    0.028830929 = product of:
      0.12973918 = sum of:
        0.028470935 = weight(_text_:data in 604) [ClassicSimilarity], result of:
          0.028470935 = score(doc=604,freq=2.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.24455236 = fieldWeight in 604, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0546875 = fieldNorm(doc=604)
        0.10126825 = weight(_text_:germany in 604) [ClassicSimilarity], result of:
          0.10126825 = score(doc=604,freq=2.0), product of:
            0.21956629 = queryWeight, product of:
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.036818076 = queryNorm
            0.46121946 = fieldWeight in 604, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.0546875 = fieldNorm(doc=604)
      0.22222222 = coord(2/9)
    
    Abstract
    iQvoc is a new open source SKOS-XL vocabulary management tool developed by the Federal Environment Agency, Germany, and innoQ Deutschland GmbH. Its immediate purpose is maintaining and publishing reference vocabularies in the upcoming Linked Data cloud of environmental information, but it may be easily adapted to host any SKOS- XL compliant vocabulary. iQvoc is implemented as a Ruby on Rails application running on top of JRuby - the Java implementation of the Ruby Programming Language. To increase the user experience when editing content, iQvoc uses heavily the JavaScript library jQuery.
  3. Manguinhas, H.; Charles, V.; Isaac, A.; Miles, T.; Lima, A.; Neroulidis, A.; Ginouves, V.; Atsidis, D.; Hildebrand, M.; Brinkerink, M.; Gordea, S.: Linking subject labels in cultural heritage metadata to MIMO vocabulary using CultuurLink (2016) 0.03
    0.02695852 = product of:
      0.12131333 = sum of:
        0.034511987 = weight(_text_:data in 3107) [ClassicSimilarity], result of:
          0.034511987 = score(doc=3107,freq=4.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.29644224 = fieldWeight in 3107, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046875 = fieldNorm(doc=3107)
        0.08680135 = weight(_text_:germany in 3107) [ClassicSimilarity], result of:
          0.08680135 = score(doc=3107,freq=2.0), product of:
            0.21956629 = queryWeight, product of:
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.036818076 = queryNorm
            0.39533097 = fieldWeight in 3107, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.046875 = fieldNorm(doc=3107)
      0.22222222 = coord(2/9)
    
    Abstract
    The Europeana Sounds project aims to increase the amount of cultural audio content in Europeana. It also strongly focuses on enriching the metadata records that are aggregated by Europeana. To provide metadata to Europeana, Data Providers are asked to convert their records from the format and model they use internally to a specific profile of the Europeana Data Model (EDM) for sound resources. These metadata include subjects, which typically use a vocabulary internal to each partner.
    Source
    Proceedings of the 15th European Networked Knowledge Organization Systems Workshop (NKOS 2016) co-located with the 20th International Conference on Theory and Practice of Digital Libraries 2016 (TPDL 2016), Hannover, Germany, September 9, 2016. Edi. by Philipp Mayr et al. [http://ceur-ws.org/Vol-1676/=urn:nbn:de:0074-1676-5]
  4. Manguinhas, H.; Freire, N.; Machado, J.; Borbinha, J.: Supporting multilingual bibliographic resource discovery with Functional Requirements for Bibliographic Records (2012) 0.03
    0.025069747 = product of:
      0.11281386 = sum of:
        0.09247748 = weight(_text_:bibliographic in 133) [ClassicSimilarity], result of:
          0.09247748 = score(doc=133,freq=18.0), product of:
            0.14333439 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.036818076 = queryNorm
            0.64518696 = fieldWeight in 133, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.0390625 = fieldNorm(doc=133)
        0.020336384 = weight(_text_:data in 133) [ClassicSimilarity], result of:
          0.020336384 = score(doc=133,freq=2.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.17468026 = fieldWeight in 133, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0390625 = fieldNorm(doc=133)
      0.22222222 = coord(2/9)
    
    Abstract
    This paper describes an experiment exploring the hypothesis that innovative application of the Functional Require-ments for Bibliographic Records (FRBR) principles can complement traditional bibliographic resource discovery systems in order to improve the user experience. A specialized service was implemented that, when given a plain list of results from a regular online catalogue, was able to process, enrich and present that list in a more relevant way for the user. This service pre-processes the records of a traditional online catalogue in order to build a semantic structure following the FRBR model. The service also explores web search features that have been revolutionizing the way users conceptualize resource discovery, such as relevance ranking and metasearching. This work was developed in the context of the TELPlus project. We processed nearly one hundred thousand bibliographic and authority records, in multiple languages, and originating from twelve European na-tional libraries. This paper describes the architecture of the service and the main challenges faced, especially concerning the extraction and linking of the relevant FRBR entities from the bibliographic metadata produced by the libraries. The service was evaluated by end users, who filled out a questionnaire after using a traditional online catalogue and the new service, both with the same bibliographic collection. The analysis of the results supports the hypothesis that FRBR can be implemented for re-source discovery in a non-intrusive way, reusing the data of any existing traditional bibliographic system.
    Content
    Beitrag eines Schwerpunktthemas: Semantic Web and Reasoning for Cultural Heritage and Digital Libraries: Vgl.: http://www.semantic-web-journal.net/content/supporting-multilingual-bibliographic-resource-discovery-functional-requirements-bibliograph http://www.semantic-web-journal.net/sites/default/files/swj145_2.pdf.
  5. McGrath, K.; Kules, B.; Fitzpatrick, C.: FRBR and facets provide flexible, work-centric access to items in library collections (2011) 0.02
    0.024521142 = product of:
      0.11034514 = sum of:
        0.061032027 = weight(_text_:bibliographic in 2430) [ClassicSimilarity], result of:
          0.061032027 = score(doc=2430,freq=4.0), product of:
            0.14333439 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.036818076 = queryNorm
            0.4258017 = fieldWeight in 2430, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2430)
        0.04931311 = weight(_text_:data in 2430) [ClassicSimilarity], result of:
          0.04931311 = score(doc=2430,freq=6.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.42357713 = fieldWeight in 2430, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2430)
      0.22222222 = coord(2/9)
    
    Abstract
    This paper explores a technique to improve searcher access to library collections by providing a faceted search interface built on a data model based on the Functional Requirements for Bibliographic Records (FRBR). The prototype provides a Workcentric view of a moving image collection that is integrated with bibliographic and holdings data. Two sets of facets address important user needs: "what do you want?" and "how/where do you want it?" enabling patrons to narrow, broaden and pivot across facet values instead of limiting them to the tree-structured hierarchy common with existing FRBR applications. The data model illustrates how FRBR is being adapted and applied beyond the traditional library catalog.
  6. Voß, J.: Classification of knowledge organization systems with Wikidata (2016) 0.02
    0.02261475 = product of:
      0.10176637 = sum of:
        0.08680135 = weight(_text_:germany in 3082) [ClassicSimilarity], result of:
          0.08680135 = score(doc=3082,freq=2.0), product of:
            0.21956629 = queryWeight, product of:
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.036818076 = queryNorm
            0.39533097 = fieldWeight in 3082, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.046875 = fieldNorm(doc=3082)
        0.014965023 = product of:
          0.029930046 = sum of:
            0.029930046 = weight(_text_:22 in 3082) [ClassicSimilarity], result of:
              0.029930046 = score(doc=3082,freq=2.0), product of:
                0.12893063 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.036818076 = queryNorm
                0.23214069 = fieldWeight in 3082, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3082)
          0.5 = coord(1/2)
      0.22222222 = coord(2/9)
    
    Pages
    S.15-22
    Source
    Proceedings of the 15th European Networked Knowledge Organization Systems Workshop (NKOS 2016) co-located with the 20th International Conference on Theory and Practice of Digital Libraries 2016 (TPDL 2016), Hannover, Germany, September 9, 2016. Edi. by Philipp Mayr et al. [http://ceur-ws.org/Vol-1676/=urn:nbn:de:0074-1676-5]
  7. Leresche, F.; Boulet, V.: RDA as a tool for the bibliographic transition : the French position (2016) 0.02
    0.02247122 = product of:
      0.10112049 = sum of:
        0.052313168 = weight(_text_:bibliographic in 2953) [ClassicSimilarity], result of:
          0.052313168 = score(doc=2953,freq=4.0), product of:
            0.14333439 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.036818076 = queryNorm
            0.3649729 = fieldWeight in 2953, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.046875 = fieldNorm(doc=2953)
        0.048807316 = weight(_text_:data in 2953) [ClassicSimilarity], result of:
          0.048807316 = score(doc=2953,freq=8.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.4192326 = fieldWeight in 2953, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046875 = fieldNorm(doc=2953)
      0.22222222 = coord(2/9)
    
    Abstract
    This article presents the process adopted by the France to bring library catalogs to the Web of data and the RDA role in this general strategy. After analising RDA limits and inconsistencies, inherited from the tradition of AACR and MARC21 catalogues, the authors present the French approach to RDA and its positioning in correlation to international standards like ISBD and FRBR. The method adopted in France for FRBRising the catalogues go through a technical work of creating alignment beteween existing data, exploiting the technologies applied to the creation of data.bnf.fr and through a revision of the French cataloguing rules, allowing FRBRised metadata creation. This revision is based on RDA and it is setting up a French RDA application profile, keeping the analysis on the greater differences. RDA adoption, actually, is not a crucial issue in France and not a self standing purpose; it is just a tool for the transition of bibliographic data towards the Web of data.
  8. Gómez-Pérez, A.; Corcho, O.: Ontology languages for the Semantic Web (2015) 0.02
    0.021580642 = product of:
      0.09711289 = sum of:
        0.076776505 = weight(_text_:readable in 3297) [ClassicSimilarity], result of:
          0.076776505 = score(doc=3297,freq=2.0), product of:
            0.2262076 = queryWeight, product of:
              6.1439276 = idf(docFreq=257, maxDocs=44218)
              0.036818076 = queryNorm
            0.33940727 = fieldWeight in 3297, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.1439276 = idf(docFreq=257, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3297)
        0.020336384 = weight(_text_:data in 3297) [ClassicSimilarity], result of:
          0.020336384 = score(doc=3297,freq=2.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.17468026 = fieldWeight in 3297, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3297)
      0.22222222 = coord(2/9)
    
    Abstract
    Ontologies have proven to be an essential element in many applications. They are used in agent systems, knowledge management systems, and e-commerce platforms. They can also generate natural language, integrate intelligent information, provide semantic-based access to the Internet, and extract information from texts in addition to being used in many other applications to explicitly declare the knowledge embedded in them. However, not only are ontologies useful for applications in which knowledge plays a key role, but they can also trigger a major change in current Web contents. This change is leading to the third generation of the Web-known as the Semantic Web-which has been defined as "the conceptual structuring of the Web in an explicit machine-readable way."1 This definition does not differ too much from the one used for defining an ontology: "An ontology is an explicit, machinereadable specification of a shared conceptualization."2 In fact, new ontology-based applications and knowledge architectures are developing for this new Web. A common claim for all of these approaches is the need for languages to represent the semantic information that this Web requires-solving the heterogeneous data exchange in this heterogeneous environment. Here, we don't decide which language is best of the Semantic Web. Rather, our goal is to help developers find the most suitable language for their representation needs. The authors analyze the most representative ontology languages created for the Web and compare them using a common framework.
  9. Campbell, D.G.; Mayhew, A.: ¬A phylogenetic approach to bibliographic families and relationships (2017) 0.02
    0.02129865 = product of:
      0.095843926 = sum of:
        0.075507544 = weight(_text_:bibliographic in 3875) [ClassicSimilarity], result of:
          0.075507544 = score(doc=3875,freq=12.0), product of:
            0.14333439 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.036818076 = queryNorm
            0.52679294 = fieldWeight in 3875, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3875)
        0.020336384 = weight(_text_:data in 3875) [ClassicSimilarity], result of:
          0.020336384 = score(doc=3875,freq=2.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.17468026 = fieldWeight in 3875, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3875)
      0.22222222 = coord(2/9)
    
    Abstract
    This presentation applies the principles of phylogenetic classification to the phenomenon of bibliographic relationships in library catalogues. We argue that while the FRBR paradigm supports hierarchical bibliographic relationships between works and their various expressions and manifestations, we need a different paradigm to support associative bibliographic relationships of the kind detected in previous research. Numerous studies have shown the existence and importance of bibliographic relationships that lie outside that hierarchical FRBR model: particularly the importance of bibliographic families. We would like to suggest phylogenetics as a potential means of gaining access to those more elusive and ephemeral relationships. Phylogenetic analysis does not follow the Platonic conception of an abstract work that gives rise to specific instantiations; rather, it tracks relationships of kinship as they evolve over time. We use two examples to suggest ways in which phylogenetic trees could be represented in future library catalogues. The novels of Jane Austen are used to indicate how phylogenetic trees can represent, with greater accuracy, the line of Jane Austen adaptations, ranging from contemporary efforts to complete her unfinished work, through to the more recent efforts to graft horror memes onto the original text. Stanley Kubrick's 2001: A Space Odyssey provides an example of charting relationships both backwards and forwards in time, across different media and genres. We suggest three possible means of applying phylogenetic s in the future: enhancement of the relationship designators in RDA, crowdsourcing user tags, and extracting relationship trees through big data analysis.
  10. Takhirov, N.; Aalberg, T.; Duchateau, F.; Zumer, M.: FRBR-ML: a FRBR-based framework for semantic interoperability (2012) 0.02
    0.021186067 = product of:
      0.0953373 = sum of:
        0.049321324 = weight(_text_:bibliographic in 134) [ClassicSimilarity], result of:
          0.049321324 = score(doc=134,freq=8.0), product of:
            0.14333439 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.036818076 = queryNorm
            0.34409973 = fieldWeight in 134, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.03125 = fieldNorm(doc=134)
        0.04601598 = weight(_text_:data in 134) [ClassicSimilarity], result of:
          0.04601598 = score(doc=134,freq=16.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.3952563 = fieldWeight in 134, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.03125 = fieldNorm(doc=134)
      0.22222222 = coord(2/9)
    
    Abstract
    Metadata related to cultural items such as literature, music and movies is a valuable resource that is currently exploited in many applications and services based on semantic web technologies. A vast amount of such information has been created by memory institutions in the last decades using different standard or ad hoc schemas, and a main challenge is to make this legacy data accessible as reusable semantic data. On one hand, this is a syntactic problem that can be solved by transforming to formats that are compatible with the tools and services used for semantic aware services. On the other hand, this is a semantic problem. Simply transforming from one format to another does not automatically enable semantic interoperability and legacy data often needs to be reinterpreted as well as transformed. The conceptual model in the Functional Requirements for Bibliographic Records, initially developed as a conceptual framework for library standards and systems, is a major step towards a shared semantic model of the products of artistic and intellectual endeavor of mankind. The model is generally accepted as sufficiently generic to serve as a conceptual framework for a broad range of cultural heritage metadata. Unfortunately, the existing large body of legacy data makes a transition to this model difficult. For instance, most bibliographic data is still only available in various MARC-based formats which is hard to render into reusable and meaningful semantic data. Making legacy bibliographic data accessible as semantic data is a complex problem that includes interpreting and transforming the information. In this article, we present our work on transforming and enhancing legacy bibliographic information into a representation where the structure and semantics of the FRBR model is explicit.
  11. Suchowolec, K.; Lang, C.; Schneider, R.: Re-designing online terminology resources for German grammar (2016) 0.02
    0.02059352 = product of:
      0.09267084 = sum of:
        0.020336384 = weight(_text_:data in 3108) [ClassicSimilarity], result of:
          0.020336384 = score(doc=3108,freq=2.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.17468026 = fieldWeight in 3108, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3108)
        0.07233446 = weight(_text_:germany in 3108) [ClassicSimilarity], result of:
          0.07233446 = score(doc=3108,freq=2.0), product of:
            0.21956629 = queryWeight, product of:
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.036818076 = queryNorm
            0.32944247 = fieldWeight in 3108, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3108)
      0.22222222 = coord(2/9)
    
    Abstract
    The compilation of terminological vocabularies plays a central role in the organization and retrieval of scientific texts. Both simple keyword lists as well as sophisticated modellings of relationships between terminological concepts can make a most valuable contribution to the analysis, classification, and finding of appropriate digital documents, either on the Web or within local repositories. This seems especially true for long-established scientific fields with various theoretical and historical branches, such as linguistics, where the use of terminology within documents from different origins is sometimes far from being consistent. In this short paper, we report on the early stages of a project that aims at the re-design of an existing domain-specific KOS for grammatical content grammis. In particular, we deal with the terminological part of grammis and present the state-of-the-art of this online resource as well as the key re-design principles. Further, we propose questions regarding ramifications of the Linked Open Data and Semantic Web approaches for our re-design decisions.
    Source
    Proceedings of the 15th European Networked Knowledge Organization Systems Workshop (NKOS 2016) co-located with the 20th International Conference on Theory and Practice of Digital Libraries 2016 (TPDL 2016), Hannover, Germany, September 9, 2016. Edi. by Philipp Mayr et al. [http://ceur-ws.org/Vol-1676/=urn:nbn:de:0074-1676-5]
  12. Petric, T.: Bibliographic organisation of continuing resources in relation to the IFLA models : research within the Croatian corpus of continuing resources (2016) 0.02
    0.019115554 = product of:
      0.08601999 = sum of:
        0.06975088 = weight(_text_:bibliographic in 2960) [ClassicSimilarity], result of:
          0.06975088 = score(doc=2960,freq=16.0), product of:
            0.14333439 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.036818076 = queryNorm
            0.4866305 = fieldWeight in 2960, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.03125 = fieldNorm(doc=2960)
        0.016269106 = weight(_text_:data in 2960) [ClassicSimilarity], result of:
          0.016269106 = score(doc=2960,freq=2.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.1397442 = fieldWeight in 2960, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.03125 = fieldNorm(doc=2960)
      0.22222222 = coord(2/9)
    
    Abstract
    Comprehensive research on continuing resources has not been conducted in Croatia, therefore this paper will indicate the current bibliographic organisation of continuing resources in comparison to the parameters set by the IFLA models, and the potential flaws of the IFLA models in the bibliographic organisation of continuing resources, in comparison to the valid national code which is used in Croatian cataloguing practice. Research on the corpus of Croatian continuing resources was performed in the period from 2000 and 2011. By using the listed population through the method of deliberate stratified sampling, the titles which had been observed were selected. Through the method of observation of bibliographic records of the selected sample in the NUL catalogue, the frequency of occurrence of parameters from the IFLA models that should identify continuing resources will be recorded and should also show the characteristics of continuing resources. In determining the parameters of observation, the FRBR model is viewed in terms of bibliographic data, FRAD is viewed in terms of other groups or entities or controlled access points for work, person and the corporate body and FRSAD in terms of the third group of entities as the subject or the subject access to continuing resources. Research results indicate that the current model of bibliographic organisation presents a high frequency of attributes that are listed in the IFLA models for all types of resources, although that was not envisaged by the PPIAK, and it is clear that the practice has moved away from the national code which does not offer solutions for all types of resources and ever more so demanding users. The current model of bibliographic organisation of the corpus of Croatian continuing resources in regards to the new IFLA model requires certain changes in order for the user to more easily access and identify continuing resources. The research results also indicate the need to update the entity expression with the attribute mode of expression, and entity manifestation with the attributes mode of issuance, as well as further consideration in terms of the bibliographic organisation of continuing resources.
  13. Buttò, S.: RDA: analyses, considerations and activities by the Central Institute for the Union Catalogue of Italian Libraries and Bibliographic Information (ICCU) (2016) 0.02
    0.018219566 = product of:
      0.081988044 = sum of:
        0.06165166 = weight(_text_:bibliographic in 2958) [ClassicSimilarity], result of:
          0.06165166 = score(doc=2958,freq=8.0), product of:
            0.14333439 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.036818076 = queryNorm
            0.43012467 = fieldWeight in 2958, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2958)
        0.020336384 = weight(_text_:data in 2958) [ClassicSimilarity], result of:
          0.020336384 = score(doc=2958,freq=2.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.17468026 = fieldWeight in 2958, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2958)
      0.22222222 = coord(2/9)
    
    Abstract
    The report aims to analyze the applicability of the Resource Description and Access (RDA) within the Italian public libraries, and also in the archives and museums in order to contribute to the discussion at international level. The Central Institute for the Union Catalogue of Italian libraries (ICCU) manages the online catalogue of the Italian libraries and the network of bibliographic services. ICCU has the institutional task of coordinating the cataloging and the documentation activities for the Italian libraries. On March 31 st 2014, the Institute signed the Agreement with the American Library Association,Publishing ALA, for the Italian translation rights of RDA, now available and published inRDAToolkit. The Italian translation has been carried out and realized by the Technical Working Group, made up of the main national and academic libraries, cultural Institutions and bibliographic agencies. The Group started working from the need of studying the new code in its textual detail, to better understand the principles, purposes, and applicability and finally its sustainability within the national context in relation to the area of the bibliographic control. At international level, starting from the publication of the Italian version of RDA and through the research carried out by ICCU and by the national Working Groups, the purpose is a more direct comparison with the experiences of the other European countries, also within EURIG international context, for an exchange of experiences aimed at strengthening the informational content of the data cataloging, with respect to history, cultural traditions and national identities of the different countries.
  14. Bastos Vieira, S.; DeBrito, M.; Mustafa El Hadi, W.; Zumer, M.: Developing imaged KOS with the FRSAD Model : a conceptual methodology (2016) 0.02
    0.017972346 = product of:
      0.08087556 = sum of:
        0.02300799 = weight(_text_:data in 3109) [ClassicSimilarity], result of:
          0.02300799 = score(doc=3109,freq=4.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.19762816 = fieldWeight in 3109, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.03125 = fieldNorm(doc=3109)
        0.057867568 = weight(_text_:germany in 3109) [ClassicSimilarity], result of:
          0.057867568 = score(doc=3109,freq=2.0), product of:
            0.21956629 = queryWeight, product of:
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.036818076 = queryNorm
            0.26355398 = fieldWeight in 3109, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.963546 = idf(docFreq=308, maxDocs=44218)
              0.03125 = fieldNorm(doc=3109)
      0.22222222 = coord(2/9)
    
    Abstract
    This proposal presents the methodology of indexing with images suggested by De Brito and Caribé (2015). The imagetic model is used as a compatible mechanism with FRSAD for a global information share and use of subject data, both within the library sector and beyond. The conceptual model of imagetic indexing shows how images are related to topics and 'key-images' are interpreted as nomens to implement the FRSAD model. Indexing with images consists of using images instead of key-words or descriptors, to represent and organize information. Implementing the imaged navigation in OPACs denotes multiple advantages derived from this rethinking the OPAC anew, since we are looking forward to sharing concepts within the subject authority data. Images, carrying linguistic objects, permeate inter-social and cultural concepts. In practice it includes translated metadata, symmetrical multilingual thesaurus, or any traditional indexing tools. iOPAC embodies efforts focused on conceptual levels as expected from librarians. Imaged interfaces are more intuitive since users do not need specific training for information retrieval, offering easier comprehension of indexing codes, larger conceptual portability of descriptors (as images), and a better interoperability between discourse codes and indexing competences affecting positively social and cultural interoperability. The imagetic methodology deploys R&D fields for more suitable interfaces taking into consideration users with specific needs such as deafness and illiteracy. This methodology arouse questions about the paradigms of the primacy of orality in information systems and pave the way to a legitimacy of multiple perspectives in document indexing by suggesting a more universal communication system based on images. Interdisciplinarity in neurosciences, linguistics and information sciences would be desirable competencies for further investigations about he nature of cognitive processes in information organization and classification while developing assistive KOS for individuals with communication problems, such autism and deafness.
    Source
    Proceedings of the 15th European Networked Knowledge Organization Systems Workshop (NKOS 2016) co-located with the 20th International Conference on Theory and Practice of Digital Libraries 2016 (TPDL 2016), Hannover, Germany, September 9, 2016. Edi. by Philipp Mayr et al. [http://ceur-ws.org/Vol-1676/=urn:nbn:de:0074-1676-5]
  15. Edmunds, J.: Roadmap to nowhere : BIBFLOW, BIBFRAME, and linked data for libraries (2017) 0.02
    0.017613193 = product of:
      0.079259366 = sum of:
        0.036990993 = weight(_text_:bibliographic in 3523) [ClassicSimilarity], result of:
          0.036990993 = score(doc=3523,freq=2.0), product of:
            0.14333439 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.036818076 = queryNorm
            0.2580748 = fieldWeight in 3523, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.046875 = fieldNorm(doc=3523)
        0.042268377 = weight(_text_:data in 3523) [ClassicSimilarity], result of:
          0.042268377 = score(doc=3523,freq=6.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.3630661 = fieldWeight in 3523, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046875 = fieldNorm(doc=3523)
      0.22222222 = coord(2/9)
    
    Abstract
    On December 12, 2016, Carl Stahmer and MacKenzie Smith presented at the CNI Members Fall Meeting about the BIBFLOW project, self-described on Twitter as "a two-year project of the UC Davis University Library and Zepheira investigating the future of library technical services." In her opening remarks, Ms. Smith, University Librarian at UC Davis, stated that one of the goals of the project was to devise a roadmap "to get from where we are today, which is kind of the 1970s with a little lipstick on it, to 2020, which is where we're going to be very soon." The notion that where libraries are today is somehow behind the times is one of the commonly heard rationales behind a move to linked data. Stated more precisely: - Libraries devote considerable time and resources to producing high-quality bibliographic metadata - This metadata is stored in unconnected silos - This metadata is in a format (MARC) that is incompatible with technologies of the emerging Semantic Web - The visibility of library metadata is diminished as a result of the two points above Are these assertions true? If yes, is linked data the solution?
  16. Edmunds, J.: Zombrary apocalypse!? : RDA, LRM, and the death of cataloging (2017) 0.02
    0.01703892 = product of:
      0.07667515 = sum of:
        0.060406037 = weight(_text_:bibliographic in 3818) [ClassicSimilarity], result of:
          0.060406037 = score(doc=3818,freq=12.0), product of:
            0.14333439 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.036818076 = queryNorm
            0.42143437 = fieldWeight in 3818, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.03125 = fieldNorm(doc=3818)
        0.016269106 = weight(_text_:data in 3818) [ClassicSimilarity], result of:
          0.016269106 = score(doc=3818,freq=2.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.1397442 = fieldWeight in 3818, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.03125 = fieldNorm(doc=3818)
      0.22222222 = coord(2/9)
    
    Abstract
    A brochure on RDA issued in 2010 includes the statements that "RDA goes beyond earlier cataloguing codes in that it provides guidelines on cataloguing digital resources and a stronger emphasis on helping users find, identify, select, and obtain the information they want. RDA also supports clustering of bibliographic records to show relationships between works and their creators. This important new feature makes users more aware of a work's different editions, translations, or physical formats - an exciting development." Setting aside the fact that the author(s) of these statements and I differ on the definition of exciting, their claims are, at best, dubious. There is no evidence-empirical or anecdotal-that bibliographic records created using RDA are any better than records created using AACR2 (or AACR, for that matter) in "helping users find, identify, select, and obtain the information they want." The claim is especially unfounded in the context of the current discovery ecosystem, in which users are perfectly capable of finding, identifying, selecting, and obtaining information with absolutely no assistance from libraries or the bibliographic data libraries create.
    Equally fallacious is the statement that support for the "clustering bibliographic records to show relationships between works and their creators" is an "important new feature" of RDA. AACR2 bibliographic records and the systems housing them can, did, and do show such relationships. Finally, whether users want or care to be made "more aware of a work's different editions, translations, or physical formats" is debatable. As an aim, it sounds less like what a user wants and more like what a cataloging librarian thinks a user should want. As Amanda Cossham writes in her recently issued doctoral thesis: "The explicit focus on user needs in the FRBR model, the International Cataloguing Principles, and RDA: Resource Description and Access does not align well with the ways that users use, understand, and experience library catalogues nor with the ways that they understand and experience the wider information environment. User tasks, as constituted in the FRBR model and RDA, are insufficient to meet users' needs." (p. 11, emphasis in the original)
    The point of this paper is not to critique RDA (a futile task, since RDA is here to stay), but to make plain that its claim to be a solution to the challenge(s) of bibliographic description in the Internet Age is unfounded, and, secondarily, to explain why such wild claims continue to be advanced and go unchallenged by the rank and file of career catalogers.
  17. Tillett, B.B.: RDA, or, The long journey of the catalog to the digital age (2016) 0.02
    0.015917132 = product of:
      0.071627095 = sum of:
        0.04315616 = weight(_text_:bibliographic in 2945) [ClassicSimilarity], result of:
          0.04315616 = score(doc=2945,freq=2.0), product of:
            0.14333439 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.036818076 = queryNorm
            0.30108726 = fieldWeight in 2945, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2945)
        0.028470935 = weight(_text_:data in 2945) [ClassicSimilarity], result of:
          0.028470935 = score(doc=2945,freq=2.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.24455236 = fieldWeight in 2945, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2945)
      0.22222222 = coord(2/9)
    
    Abstract
    RDA was created in response to complaints about the Anglo-American Cataloguing Rules, especially the call for a more international, principle-based content standard that takes the perspective of the conceptual models of FRBR (Functional Requirements for Bibliographic Records) and FRAD (Functional Requirements for Authority Data). The past and ongoing process for continuous improvement to RDA is through the Joint Steering Committee for Development of RDA (known as the JSC, but recently renamed the RDA Steering Committee - RSC) to make RDA even more international and principle-based.
  18. Forero, D.; Peterson, N.; Hamilton, A.: Building an institutional author search tool (2019) 0.02
    0.015917132 = product of:
      0.071627095 = sum of:
        0.04315616 = weight(_text_:bibliographic in 5441) [ClassicSimilarity], result of:
          0.04315616 = score(doc=5441,freq=2.0), product of:
            0.14333439 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.036818076 = queryNorm
            0.30108726 = fieldWeight in 5441, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5441)
        0.028470935 = weight(_text_:data in 5441) [ClassicSimilarity], result of:
          0.028470935 = score(doc=5441,freq=2.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.24455236 = fieldWeight in 5441, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5441)
      0.22222222 = coord(2/9)
    
    Abstract
    Ability to collect time-specific lists of faculty publications has become increasingly important for academic departments. At OHSU publication lists had been retrieved manually by a librarian who conducted literature searches in bibliographic databases. These searches were complicated and time consuming, and the results were large and difficult to assess for accuracy. The OHSU library has built an open web page that allows novices to make very sophisticated institution-specific queries. The tool frees up library staff, provides users with an easy way of retrieving reliable local publication information from PubMed, and gives an opportunity for more sophisticated users to modify the algorithm or dive into the data to better understand nuances from a strong jumping off point.
  19. Mitchell, J.S.; Zeng, M.L.; Zumer, M.: Modeling classification systems in multicultural and multilingual contexts (2012) 0.02
    0.015549123 = product of:
      0.069971055 = sum of:
        0.048807316 = weight(_text_:data in 1967) [ClassicSimilarity], result of:
          0.048807316 = score(doc=1967,freq=8.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.4192326 = fieldWeight in 1967, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046875 = fieldNorm(doc=1967)
        0.02116374 = product of:
          0.04232748 = sum of:
            0.04232748 = weight(_text_:22 in 1967) [ClassicSimilarity], result of:
              0.04232748 = score(doc=1967,freq=4.0), product of:
                0.12893063 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.036818076 = queryNorm
                0.32829654 = fieldWeight in 1967, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1967)
          0.5 = coord(1/2)
      0.22222222 = coord(2/9)
    
    Abstract
    This paper reports on the second part of an initiative of the authors on researching classification systems with the conceptual model defined by the Functional Requirements for Subject Authority Data (FRSAD) final report. In an earlier study, the authors explored whether the FRSAD conceptual model could be extended beyond subject authority data to model classification data. The focus of the current study is to determine if classification data modeled using FRSAD can be used to solve real-world discovery problems in multicultural and multilingual contexts. The paper discusses the relationships between entities (same type or different types) in the context of classification systems that involve multiple translations and /or multicultural implementations. Results of two case studies are presented in detail: (a) two instances of the DDC (DDC 22 in English, and the Swedish-English mixed translation of DDC 22), and (b) Chinese Library Classification. The use cases of conceptual models in practice are also discussed.
  20. Dousa, T.M.: E. Wyndham Hulme's classification of the attributes of books : On an early model of a core bibliographical entity (2017) 0.01
    0.01457565 = product of:
      0.06559043 = sum of:
        0.049321324 = weight(_text_:bibliographic in 3859) [ClassicSimilarity], result of:
          0.049321324 = score(doc=3859,freq=8.0), product of:
            0.14333439 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.036818076 = queryNorm
            0.34409973 = fieldWeight in 3859, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.03125 = fieldNorm(doc=3859)
        0.016269106 = weight(_text_:data in 3859) [ClassicSimilarity], result of:
          0.016269106 = score(doc=3859,freq=2.0), product of:
            0.11642061 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.036818076 = queryNorm
            0.1397442 = fieldWeight in 3859, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.03125 = fieldNorm(doc=3859)
      0.22222222 = coord(2/9)
    
    Abstract
    Modelling bibliographical entities is a prominent activity within knowledge organization today. Current models of bibliographic entities, such as Functional Requirements for Bibliographical Records (FRBR) and the Bibliographic Framework (BIBFRAME), take inspiration from data - modelling methods developed by computer scientists from the mid - 1970s on. Thus, it would seem that the modelling of bibliographic entities is an activity of very recent vintage. However, it is possible to find examples of bibliographical models from earlier periods of knowledge organization. The purpose of this paper is to draw attention to one such model, outlined by the early 20th - century British classification theorist E. Wyndham Hulme in his essay on "Principles of Book Classification" (1911 - 1912). There, Hulme set forth a classification of various attributes by which books can conceivably be classified. These he first divided into accidental and inseparable attributes. Accidental attributes were subdivided into edition - level and copy - level attributes and inseparable attitudes, into physical and non - physical attributes. Comparison of Hulme's classification of attributes with those of FRBR and BIBFRAME 2.0 reveals that the different classes of attributes in Hulme's classification correspond to groups of attributes associated with different bibliographical entities in those models. These later models assume the existence of different bibliographic entities in an abstraction hierarchy among which attributes are distributed, whereas Hulme posited only a single entity - the book - , whose various aspects he clustered into different classes of attributes. Thus, Hulme's model offers an interesting alternative to current assumptions about how to conceptualize the relationship between attributes and entities in the bibliographical universe.

Languages

  • e 97
  • d 48
  • i 3
  • a 1
  • f 1
  • More… Less…