Search (180 results, page 2 of 9)

  • × language_ss:"e"
  • × year_i:[2020 TO 2030}
  1. Binding, C.; Gnoli, C.; Tudhope, D.: Migrating a complex classification scheme to the semantic web : expressing the Integrative Levels Classification using SKOS RDF (2021) 0.00
    0.0036409218 = product of:
      0.0509729 = sum of:
        0.0509729 = weight(_text_:representation in 600) [ClassicSimilarity], result of:
          0.0509729 = score(doc=600,freq=6.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.44023782 = fieldWeight in 600, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.0390625 = fieldNorm(doc=600)
      0.071428575 = coord(1/14)
    
    Abstract
    Purpose The Integrative Levels Classification (ILC) is a comprehensive "freely faceted" knowledge organization system not previously expressed as SKOS (Simple Knowledge Organization System). This paper reports and reflects on work converting the ILC to SKOS representation. Design/methodology/approach The design of the ILC representation and the various steps in the conversion to SKOS are described and located within the context of previous work considering the representation of complex classification schemes in SKOS. Various issues and trade-offs emerging from the conversion are discussed. The conversion implementation employed the STELETO transformation tool. Findings The ILC conversion captures some of the ILC facet structure by a limited extension beyond the SKOS standard. SPARQL examples illustrate how this extension could be used to create faceted, compound descriptors when indexing or cataloguing. Basic query patterns are provided that might underpin search systems. Possible routes for reducing complexity are discussed. Originality/value Complex classification schemes, such as the ILC, have features which are not straight forward to represent in SKOS and which extend beyond the functionality of the SKOS standard. The ILC's facet indicators are modelled as rdf:Property sub-hierarchies that accompany the SKOS RDF statements. The ILC's top-level fundamental facet relationships are modelled by extensions of the associative relationship - specialised sub-properties of skos:related. An approach for representing faceted compound descriptions in ILC and other faceted classification schemes is proposed.
  2. Tan, X.; Luo, X.; Wang, X.; Wang, H.; Hou, X.: Representation and display of digital images of cultural heritage : a semantic enrichment approach (2021) 0.00
    0.0035673599 = product of:
      0.049943037 = sum of:
        0.049943037 = weight(_text_:representation in 455) [ClassicSimilarity], result of:
          0.049943037 = score(doc=455,freq=4.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.4313432 = fieldWeight in 455, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.046875 = fieldNorm(doc=455)
      0.071428575 = coord(1/14)
    
    Abstract
    Digital images of cultural heritage (CH) contain rich semantic information. However, today's semantic representations of CH images fail to fully reveal the content entities and context within these vital surrogates. This paper draws on the fields of image research and digital humanities to propose a systematic methodology and a technical route for semantic enrichment of CH digital images. This new methodology systematically applies a series of procedures including: semantic annotation, entity-based enrichment, establishing internal relations, event-centric enrichment, defining hierarchy relations between properties text annotation, and finally, named entity recognition in order to ultimately provide fine-grained contextual semantic content disclosure. The feasibility and advantages of the proposed semantic enrichment methods for semantic representation are demonstrated via a visual display platform for digital images of CH built to represent the Wutai Mountain Map, a typical Dunhuang mural. This study proves that semantic enrichment offers a promising new model for exposing content at a fine-grained level, and establishing a rich semantic network centered on the content of digital images of CH.
  3. Song, N.; Cheng, H.; Zhou, H.; Wang, X.: Linking scholarly contents : the design and construction of an argumentation graph (2022) 0.00
    0.0035673599 = product of:
      0.049943037 = sum of:
        0.049943037 = weight(_text_:representation in 1104) [ClassicSimilarity], result of:
          0.049943037 = score(doc=1104,freq=4.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.4313432 = fieldWeight in 1104, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.046875 = fieldNorm(doc=1104)
      0.071428575 = coord(1/14)
    
    Abstract
    In this study, we propose a way to link the scholarly contents of scientific papers by constructing a knowledge graph based on the semantic organization of argumentation units and relations in scientific papers. We carried out an argumentation graph data model aimed at linking multiple discourses, and also developed a semantic annotation platform for scientific papers and an argumentation graph visualization system. A construction experiment was performed using 12 articles. The final argumentation graph has 1,262 nodes and 1,628 edges, including 1,628 intra-article relations and 190 inter-article relations. Knowledge evolution representation, strategic reading, and automatic abstracting use cases are presented to demonstrate the application of the argumentation graph. In contrast to existing knowledge graphs used in academic fields, the argumentation graph better supports the organization and representation of scientific paper content and can be used as data infrastructure in scientific knowledge retrieval, reorganization, reasoning, and evolution. Moreover, it supports automatic abstract and strategic reading.
  4. Zavalin, V.: Exploration of subject and genre representation in bibliographic metadata representing works of fiction for children and young adults (2024) 0.00
    0.0035673599 = product of:
      0.049943037 = sum of:
        0.049943037 = weight(_text_:representation in 1152) [ClassicSimilarity], result of:
          0.049943037 = score(doc=1152,freq=4.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.4313432 = fieldWeight in 1152, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.046875 = fieldNorm(doc=1152)
      0.071428575 = coord(1/14)
    
    Abstract
    This study examines subject and genre representation in metadata that describes information resources created for children and young adult audiences. Both quantitative and limited qualitative analyses were applied to the analysis of WorldCat records collected in 2021 and contributed by the Children's and Young Adults' Cataloging Program at the US Library of Congress. This dataset contains records created several years prior to the data collection point and edited by various OCLC member institutions. Findings provide information on the level and patterns of application of these kinds of metadata important for information access, with a focus on the fields, subfields, and controlled vocabularies used. The discussion of results includes a detailed evaluation of genre and subject metadata quality (accuracy, completeness, and consistency).
  5. Robertson, C.: ¬The filing cabinet : a vertical history of information (2021) 0.00
    0.0033895585 = product of:
      0.047453817 = sum of:
        0.047453817 = weight(_text_:mental in 641) [ClassicSimilarity], result of:
          0.047453817 = score(doc=641,freq=2.0), product of:
            0.16438161 = queryWeight, product of:
              6.532101 = idf(docFreq=174, maxDocs=44218)
              0.025165197 = queryNorm
            0.28868082 = fieldWeight in 641, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.532101 = idf(docFreq=174, maxDocs=44218)
              0.03125 = fieldNorm(doc=641)
      0.071428575 = coord(1/14)
    
    Abstract
    The ubiquity of the filing cabinet in the twentieth-century office space, along with its noticeable absence of style, has obscured its transformative role in the histories of both information technology and work. In the first in-depth history of this neglected artifact, Craig Robertson explores how the filing cabinet profoundly shaped the way that information and data have been sorted, stored, retrieved, and used. Invented in the 1890s, the filing cabinet was a result of the nineteenth-century faith in efficiency. Previously, paper records were arranged haphazardly: bound into books, stacked in piles, curled into slots, or impaled on spindles. The filing cabinet organized loose papers in tabbed folders that could be sorted alphanumerically, radically changing how people accessed, circulated, and structured information. Robertson's unconventional history of the origins of the information age posits the filing cabinet as an information storage container, an 'automatic memory' machine that contributed to a new type of information labor privileging manual dexterity over mental deliberation. Gendered assumptions about women's nimble fingers helped to naturalize the changes that brought women into the workforce as low-level clerical workers. The filing cabinet emerges from this unexpected account as a sophisticated piece of information technology and a site of gendered labor that with its folders, files, and tabs continues to shape how we interact with information and data in today's digital world.
  6. Gil-Berrozpe, J.C.: Description, categorization, and representation of hyponymy in environmental terminology (2022) 0.00
    0.0033633395 = product of:
      0.04708675 = sum of:
        0.04708675 = weight(_text_:representation in 1004) [ClassicSimilarity], result of:
          0.04708675 = score(doc=1004,freq=8.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.40667427 = fieldWeight in 1004, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.03125 = fieldNorm(doc=1004)
      0.071428575 = coord(1/14)
    
    Abstract
    Terminology has evolved from static and prescriptive theories to dynamic and cognitive approaches. Thanks to these approaches, there have been significant advances in the design and elaboration of terminological resources. This has resulted in the creation of tools such as terminological knowledge bases, which are able to show how concepts are interrelated through different semantic or conceptual relations. Of these relations, hyponymy is the most relevant to terminology work because it deals with concept categorization and term hierarchies. This doctoral thesis presents an enhancement of the semantic structure of EcoLexicon, a terminological knowledge base on environmental science. The aim of this research was to improve the description, categorization, and representation of hyponymy in environmental terminology. Therefore, we created HypoLexicon, a new stand-alone module for EcoLexicon in the form of a hyponymy-based terminological resource. This resource contains twelve terminological entries from four specialized domains (Biology, Chemistry, Civil Engineering, and Geology), which consist of 309 concepts and 465 terms associated with those concepts. This research was mainly based on the theoretical premises of Frame-based Terminology. This theory was combined with Cognitive Linguistics, for conceptual description and representation; Corpus Linguistics, for the extraction and processing of linguistic and terminological information; and Ontology, related to hyponymy and relevant for concept categorization. HypoLexicon was constructed from the following materials: (i) the EcoLexicon English Corpus; (ii) other specialized terminological resources, including EcoLexicon; (iii) Sketch Engine; and (iv) Lexonomy. This thesis explains the methodologies applied for corpus extraction and compilation, corpus analysis, the creation of conceptual hierarchies, and the design of the terminological template. The results of the creation of HypoLexicon are discussed by highlighting the information in the hyponymy-based terminological entries: (i) parent concept (hypernym); (ii) child concepts (hyponyms, with various hyponymy levels); (iii) terminological definitions; (iv) conceptual categories; (v) hyponymy subtypes; and (vi) hyponymic contexts. Furthermore, the features and the navigation within HypoLexicon are described from the user interface and the admin interface. In conclusion, this doctoral thesis lays the groundwork for developing a terminological resource that includes definitional, relational, ontological and contextual information about specialized hypernyms and hyponyms. All of this information on specialized knowledge is simple to follow thanks to the hierarchical structure of the terminological template used in HypoLexicon. Therefore, not only does it enhance knowledge representation, but it also facilitates its acquisition.
  7. Escolano, C.; Costa-Jussà, M.R.; Fonollosa, J.A.: From bilingual to multilingual neural-based machine translation by incremental training (2021) 0.00
    0.0029727998 = product of:
      0.041619197 = sum of:
        0.041619197 = weight(_text_:representation in 97) [ClassicSimilarity], result of:
          0.041619197 = score(doc=97,freq=4.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.35945266 = fieldWeight in 97, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.0390625 = fieldNorm(doc=97)
      0.071428575 = coord(1/14)
    
    Abstract
    A common intermediate language representation in neural machine translation can be used to extend bilingual systems by incremental training. We propose a new architecture based on introducing an interlingual loss as an additional training objective. By adding and forcing this interlingual loss, we can train multiple encoders and decoders for each language, sharing among them a common intermediate representation. Translation results on the low-resource tasks (Turkish-English and Kazakh-English tasks) show a BLEU improvement of up to 2.8 points. However, results on a larger dataset (Russian-English and Kazakh-English) show BLEU losses of a similar amount. While our system provides improvements only for the low-resource tasks in terms of translation quality, our system is capable of quickly deploying new language pairs without the need to retrain the rest of the system, which may be a game changer in some situations. Specifically, what is most relevant regarding our architecture is that it is capable of: reducing the number of production systems, with respect to the number of languages, from quadratic to linear; incrementally adding a new language to the system without retraining the languages already there; and allowing for translations from the new language to all the others present in the system.
  8. Auer, S.; Oelen, A.; Haris, A.M.; Stocker, M.; D'Souza, J.; Farfar, K.E.; Vogt, L.; Prinz, M.; Wiens, V.; Jaradeh, M.Y.: Improving access to scientific literature with knowledge graphs : an experiment using library guidelines to judge information integrity (2020) 0.00
    0.0029727998 = product of:
      0.041619197 = sum of:
        0.041619197 = weight(_text_:representation in 316) [ClassicSimilarity], result of:
          0.041619197 = score(doc=316,freq=4.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.35945266 = fieldWeight in 316, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.0390625 = fieldNorm(doc=316)
      0.071428575 = coord(1/14)
    
    Abstract
    The transfer of knowledge has not changed fundamentally for many hundreds of years: It is usually document-based-formerly printed on paper as a classic essay and nowadays as PDF. With around 2.5 million new research contributions every year, researchers drown in a flood of pseudo-digitized PDF publications. As a result research is seriously weakened. In this article, we argue for representing scholarly contributions in a structured and semantic way as a knowledge graph. The advantage is that information represented in a knowledge graph is readable by machines and humans. As an example, we give an overview on the Open Research Knowledge Graph (ORKG), a service implementing this approach. For creating the knowledge graph representation, we rely on a mixture of manual (crowd/expert sourcing) and (semi-)automated techniques. Only with such a combination of human and machine intelligence, we can achieve the required quality of the representation to allow for novel exploration and assistance services for researchers. As a result, a scholarly knowledge graph such as the ORKG can be used to give a condensed overview on the state-of-the-art addressing a particular research quest, for example as a tabular comparison of contributions according to various characteristics of the approaches. Further possible intuitive access interfaces to such scholarly knowledge graphs include domain-specific (chart) visualizations or answering of natural language questions.
  9. Wang, X.; Song, N.; Zhou, H.; Cheng, H.: ¬The representation of argumentation in scientific papers : a comparative analysis of two research areas (2022) 0.00
    0.0029727998 = product of:
      0.041619197 = sum of:
        0.041619197 = weight(_text_:representation in 567) [ClassicSimilarity], result of:
          0.041619197 = score(doc=567,freq=4.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.35945266 = fieldWeight in 567, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.0390625 = fieldNorm(doc=567)
      0.071428575 = coord(1/14)
    
    Abstract
    Scientific papers are essential manifestations of evolving scientific knowledge, and arguments are an important avenue to communicate research results. This study aims to understand how the argumentation process is represented in scientific papers, which is important for knowledge representation, discovery, and retrieval. First, based on fundamental argument theory and scientific discourse ontologies, a coding schema, including 17 categories was constructed. Thereafter, annotation experiments were conducted with 40 scientific articles randomly selected from two different research areas (library and information science and biomedical sciences). Statistical analysis and the sequential pattern mining method were then employed; the ratio of different argumentation units and evidence types were calculated, the argumentation semantics of figures and tables analyzed, and the argumentation structures extracted. A correlation analysis between argumentation and rhetorical structures was also performed to further reveal how argumentation was represented within scientific discourses. The results indicated a difference in the proportion of the argumentation units in the two types of scientific papers, as well as a similar linear construction with differences in the specific argument structures of each knowledge domain and a clear correlation between argumentation and rhetorical structure.
  10. Lopes Martins, D.; Silva Lemos, D.L. da; Rosa de Oliveira, L.F.; Siqueira, J.; Carmo, D. do; Nunes Medeiros, V.: Information organization and representation in digital cultural heritage in Brazil : systematic mapping of information infrastructure in digital collections for data science applications (2023) 0.00
    0.0029727998 = product of:
      0.041619197 = sum of:
        0.041619197 = weight(_text_:representation in 968) [ClassicSimilarity], result of:
          0.041619197 = score(doc=968,freq=4.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.35945266 = fieldWeight in 968, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.0390625 = fieldNorm(doc=968)
      0.071428575 = coord(1/14)
    
    Abstract
    This paper focuses on data science in digital cultural heritage in Brazil, where there is a lack of systematized information and curated databases for the integrated organization of documentary knowledge. Thus, the aim was to systematically map the different forms of information organization and representation applied to objects from collections belonging to institutions affiliated with the federal government's Special Department of Culture. This diagnosis is then used to discuss the requirements of devising strategies that favor a better data science information infrastructure to reuse information on Brazil's cultural heritage. Content analysis was used to identify analytical categories and obtain a broader understanding of the documentary sources of these institutions in order to extract, analyze, and interpret the data involved. A total of 215 hyperlinks that can be considered cultural collections of the institutions studied were identified, representing 2,537,921 cultural heritage items. The results show that the online publication of Brazil's digital cultural heritage is limited in terms of technology, copyright licensing, and established information organization practices. This paper provides a conceptual and analytical view to discuss the requirements for formulating strategies aimed at building a data science information infrastructure of Brazilian digital cultural collections that serves as future projects.
  11. Taniguchi, S.: Data provenance and administrative information in library linked data : reviewing RDA in RDF, BIBFRAME, and Wikidata (2024) 0.00
    0.0029429218 = product of:
      0.041200902 = sum of:
        0.041200902 = weight(_text_:representation in 1154) [ClassicSimilarity], result of:
          0.041200902 = score(doc=1154,freq=2.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.35583997 = fieldWeight in 1154, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1154)
      0.071428575 = coord(1/14)
    
    Abstract
    We examined how data provenance and additional information of element values including nomens, and administrative information on the metadata should be modeled and represented in the Resource Description Framework (RDF) for linked data of library catalogs. First, we classified such information types into categories and organized the combination with recording-units, i.e., a description statement or description set. Next, we listed the appropriate RDF representation patterns for each recording-unit. Then, we reviewed the methods to examine such information in Resource Description and Access (RDA) in RDF, BIBFRAME, and Wikidata, and pointed out the issues involved in them.
  12. Luyt, L.: Representation and the problem of bibliographic imagination on Wikipedia (2022) 0.00
    0.0029127372 = product of:
      0.04077832 = sum of:
        0.04077832 = weight(_text_:representation in 681) [ClassicSimilarity], result of:
          0.04077832 = score(doc=681,freq=6.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.35219026 = fieldWeight in 681, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.03125 = fieldNorm(doc=681)
      0.071428575 = coord(1/14)
    
    Abstract
    Purpose The astonishing thing about Wikipedia is that despite the way it is produced, the product is as good as it is and not far worse. But this is no reason for complacency. As others have documented, Wikipedia has representational blind spots, produced by the nature of its editorial community and their discursive conventions. This article wishes to look at the potential effect of sources on certain of these blind spots. Design/methodology/approach The author used an extended example, the Wikipedia article on the Philippine-American War, to illustrate the unfortunate effects that accompany a lack of attention to the kind of sources used to produce narratives for the online encyclopaedia. The Philippine-American War article was chosen because of its importance to American history. The war brought to the fore a debate over the future of the USA and the legitimacy of a republic acquiring overseas colonies. It remains controversial today, making it essential that its representation on Wikipedia is soundly constructed. Findings Inattention to sources (a lack of bibliographical imagination) produces representational anomalies. Certain sources are privileged when they should not be and others are ignored or considered as sub-standard. Overall, the epistemological boundaries of the article in terms of what the editorial community considers reliable and what the community of scholars producing knowledge about the war think as reliable do not overlap to the extent that they should. The resulting narrative is therefore less rich than it otherwise could be. Originality/value While there exists a growing literature on the representational blind spots of Wikipedia (gender, class, geographical region and so on), the focus has been on the composition of the demographics of the editorial community. But equally important to the problem of representation are the sources used by that community. Much literature has been written that seeks to portray the social world of the marginalized, but it is not used on Wikipedia, despite it easily meeting the criteria for reliability set by the Wikipedia community. This is a tragic oversight that makes Wikipedia's aim to be a repository for the knowledge of the world, a laudable goal to strive for, even if in reality unobtainable, even harder to achieve than ever.
  13. Simoes, G.; Machado, L.; Gnoli, C.; Souza, R.: Can an ontologically-oriented KO do without concepts? (2020) 0.00
    0.0025225044 = product of:
      0.03531506 = sum of:
        0.03531506 = weight(_text_:representation in 4964) [ClassicSimilarity], result of:
          0.03531506 = score(doc=4964,freq=2.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.3050057 = fieldWeight in 4964, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.046875 = fieldNorm(doc=4964)
      0.071428575 = coord(1/14)
    
    Abstract
    The ontological approach in the development of KOS is an attempt to overcome the limitations of the traditional epistemological approach. Questions raise about the representation and organization of ontologically-oriented KO units, such as BFO universals or ILC phenomena. The study aims to compare the ontological approaches of BFO and ILC using a hermeneutic approach. We found that the differences between the units of the two systems are primarily due to the formal level of abstraction of BFO and the different organizations, namely the grouping of phenomena into ILC classes that represent complex compounds of entities in the BFO approach. In both systems the use of concepts is considered instrumental, although in the ILC they constitute the intersubjective component of the phenomena whereas in BFO they serve to access the entities of reality but are not part of them.
  14. Rocha Souza, R.; Lemos, D.: a comparative analysis : Knowledge organization systems for the representation of multimedia resources on the Web (2020) 0.00
    0.0025225044 = product of:
      0.03531506 = sum of:
        0.03531506 = weight(_text_:representation in 5993) [ClassicSimilarity], result of:
          0.03531506 = score(doc=5993,freq=2.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.3050057 = fieldWeight in 5993, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.046875 = fieldNorm(doc=5993)
      0.071428575 = coord(1/14)
    
  15. Dumitrescu, A.; Santini, S.: Full coverage of a reader's interests in context-based information filtering (2021) 0.00
    0.0025225044 = product of:
      0.03531506 = sum of:
        0.03531506 = weight(_text_:representation in 327) [ClassicSimilarity], result of:
          0.03531506 = score(doc=327,freq=2.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.3050057 = fieldWeight in 327, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.046875 = fieldNorm(doc=327)
      0.071428575 = coord(1/14)
    
    Abstract
    We present a collection of algorithms to filter a stream of documents in such a way that the filtered documents will cover as well as possible the interest of a person, keeping in mind that, at any given time, the offered documents should not only be relevant, but should also be diversified, in the sense of covering all the interests of the person. We use a modification of the WEBSOM algorithm to create a user model based on a self-organizing network trained using a collection of documents representative of the person's interests. We introduce the concepts of freshness and coverage. A document is fresh if it belongs to a semantic area of interest to a person for which no documents were seen in the recent past; a group of documents has coverage to the extent to which it is a good representation of all the interests of a person. Our tests show that these algorithms can effectively increase the coverage of the documents that are shown to the user without overly affecting precision.
  16. Wu, C.; Yan, E.; Zhu, Y.; Li, K.: Gender imbalance in the productivity of funded projects : a study of the outputs of National Institutes of Health R01 grants (2021) 0.00
    0.0025225044 = product of:
      0.03531506 = sum of:
        0.03531506 = weight(_text_:representation in 391) [ClassicSimilarity], result of:
          0.03531506 = score(doc=391,freq=2.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.3050057 = fieldWeight in 391, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.046875 = fieldNorm(doc=391)
      0.071428575 = coord(1/14)
    
    Abstract
    This study examines the relationship between team's gender composition and outputs of funded projects using a large data set of National Institutes of Health (NIH) R01 grants and their associated publications between 1990 and 2017. This study finds that while the women investigators' presence in NIH grants is generally low, higher women investigator presence is on average related to slightly lower number of publications. This study finds empirically that women investigators elect to work in fields in which fewer publications per million-dollar funding is the norm. For fields where women investigators are relatively well represented, they are as productive as men. The overall lower productivity of women investigators may be attributed to the low representation of women in high productivity fields dominated by men investigators. The findings shed light on possible reasons for gender disparity in grant productivity.
  17. Gnoli, C.: Faceted classifications as linked data : a logical analysis (2021) 0.00
    0.0025225044 = product of:
      0.03531506 = sum of:
        0.03531506 = weight(_text_:representation in 452) [ClassicSimilarity], result of:
          0.03531506 = score(doc=452,freq=2.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.3050057 = fieldWeight in 452, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.046875 = fieldNorm(doc=452)
      0.071428575 = coord(1/14)
    
    Abstract
    Faceted knowledge organization systems have sophisticated logical structures, making their representation as linked data a demanding task. The term facet is often used in ambiguous ways: while in thesauri facets only work as semantic categories, in classification schemes they also have syntactic functions. The need to convert the Integrative Levels Classification (ILC) into SKOS stimulated a more general analysis of the different kinds of syntactic facets, as can be represented in terms of RDF properties and their respective domain and range. A nomenclature is proposed, distinguishing between common facets, which can be appended to any class, that is, have an unrestricted domain; and special facets, which are exclusive to some class, that is, have a restricted domain. In both cases, foci can be taken from any other class (unrestricted range: free facets), or only from subclasses of an existing class (parallel facets), or be defined specifically for the present class (bound facets). Examples are given of such cases in ILC and in the Dewey Decimal Classification (DDC).
  18. Zhitomirsky-Geffet, M.; Avidan, G.: ¬A new framework for systematic analysis and classification of inconsistencies in multi-viewpoint ontologies (2021) 0.00
    0.0025225044 = product of:
      0.03531506 = sum of:
        0.03531506 = weight(_text_:representation in 589) [ClassicSimilarity], result of:
          0.03531506 = score(doc=589,freq=2.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.3050057 = fieldWeight in 589, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.046875 = fieldNorm(doc=589)
      0.071428575 = coord(1/14)
    
    Abstract
    Plurality of beliefs and theories in different knowledge domains calls for modelling multi-viewpoint ontologies and knowledge organization systems (KOS). A generic theoretical approach recently proposed for heterogeneity representation in KOS was linking each ontological statement to a specific validity scope to determine a set of conditions under which the statement is valid. However, the practical applicability of this approach has yet to be empirically assessed. In addition, there is still a need to investigate the types of inconsistencies that might arise in multi-viewpoint ontologies as well as their possible causes. This study proposes a new framework for systematic analysis and classification of inconsistencies in multi-viewpoint ontologies. The framework is based on eight generic logical structures of ontological statements. To test the validity of the proposed framework, two ontologies from different knowledge domains were examined. We found that only three of the eight structures led to inconsistencies in both ontologies, while the other two structures were always present in logically consistent statements. The study has practical implications for building diversified and personalized knowledge systems.
  19. Lardera, M.; Hjoerland, B.: Keyword (2021) 0.00
    0.0025225044 = product of:
      0.03531506 = sum of:
        0.03531506 = weight(_text_:representation in 591) [ClassicSimilarity], result of:
          0.03531506 = score(doc=591,freq=2.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.3050057 = fieldWeight in 591, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.046875 = fieldNorm(doc=591)
      0.071428575 = coord(1/14)
    
    Abstract
    This article discusses the different meanings of 'keyword' and related terms such as 'keyphrase', 'descriptor', 'index term', 'subject heading', 'tag' and 'n-gram' and suggests definitions of each of these terms. It further illustrates a classification of keywords, based on how they are produced or who is the actor generating them and present comparison between author-assigned keywords, indexer-assigned keywords and reader-assigned keywords as well as the automatic generation of keywords. The article also considers the functions of keywords including the use of keywords for generating bibliographic indexes. The theoretical view informing the article is that the assignment of a keyword to a text, picture or other document involves an interpretation of the document and an evaluation of the document's potentials for users. This perspective is important for both manually assigned keywords and for automated generation and is opposed to a strong tendency to consider a set of keywords as ideally presenting one best representation of a document for all requests.
  20. Bossaller, J.; Million, A.J.: ¬The research data life cycle, legacy data, and dilemmas in research data management (2023) 0.00
    0.0025225044 = product of:
      0.03531506 = sum of:
        0.03531506 = weight(_text_:representation in 966) [ClassicSimilarity], result of:
          0.03531506 = score(doc=966,freq=2.0), product of:
            0.11578492 = queryWeight, product of:
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.025165197 = queryNorm
            0.3050057 = fieldWeight in 966, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.600994 = idf(docFreq=1206, maxDocs=44218)
              0.046875 = fieldNorm(doc=966)
      0.071428575 = coord(1/14)
    
    Abstract
    This paper presents findings from an interview study of research data managers in academic data archives. Our study examined policies and professional autonomy with a focus on dilemmas encountered in everyday work by data managers. We found that dilemmas arose at every stage of the research data lifecycle, and legacy data presents particularly vexing challenges. The iFields' emphasis on knowledge organization and representation provides insight into how data, used by scientists, are used to create knowledge. The iFields' disciplinary emphasis also encompasses the sociotechnical complexity of dilemmas that we found arise in research data management. Therefore, we posit that iSchools are positioned to contribute to data science education by teaching about ethics and infrastructure used to collect, organize, and disseminate data through problem-based learning.

Types

  • a 172
  • el 9
  • m 4
  • p 4
  • x 1
  • More… Less…