Search (189 results, page 1 of 10)

  • × year_i:[2020 TO 2030}
  1. Bullard, J.; Dierking, A.; Grundner, A.: Centring LGBT2QIA+ subjects in knowledge organization systems (2020) 0.05
    0.045783717 = product of:
      0.091567434 = sum of:
        0.091567434 = sum of:
          0.054574184 = weight(_text_:subject in 5996) [ClassicSimilarity], result of:
            0.054574184 = score(doc=5996,freq=4.0), product of:
              0.16275941 = queryWeight, product of:
                3.576596 = idf(docFreq=3361, maxDocs=44218)
                0.04550679 = queryNorm
              0.33530587 = fieldWeight in 5996, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.576596 = idf(docFreq=3361, maxDocs=44218)
                0.046875 = fieldNorm(doc=5996)
          0.03699325 = weight(_text_:22 in 5996) [ClassicSimilarity], result of:
            0.03699325 = score(doc=5996,freq=2.0), product of:
              0.15935703 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04550679 = queryNorm
              0.23214069 = fieldWeight in 5996, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=5996)
      0.5 = coord(1/2)
    
    Abstract
    This paper contains a report of two interdependent knowledge organization (KO) projects for an LGBT2QIA+ library. The authors, in the context of volunteer library work for an independent library, redesigned the classification system and subject cataloguing guidelines to centre LGBT2QIA+ subjects. We discuss the priorities of creating and maintaining knowledge organization systems for a historically marginalized community and address the challenge that queer subjectivity poses to the goals of KO. The classification system features a focus on identity and physically reorganizes the library space in a way that accounts for the multiple and overlapping labels that constitute the currently articulated boundaries of this community. The subject heading system focuses on making visible topics and elements of identity made invisible by universal systems and by the newly implemented classification system. We discuss how this project may inform KO for other marginalized subjects, particularly through process and documentation that prioritizes transparency and the acceptance of an unfinished endpoint for queer KO.
    Date
    6.10.2020 21:22:33
  2. Cheti, A.; Viti, E.: Functionality and merits of a faceted thesaurus : the case of the Nuovo soggettario (2023) 0.05
    0.045783717 = product of:
      0.091567434 = sum of:
        0.091567434 = sum of:
          0.054574184 = weight(_text_:subject in 1181) [ClassicSimilarity], result of:
            0.054574184 = score(doc=1181,freq=4.0), product of:
              0.16275941 = queryWeight, product of:
                3.576596 = idf(docFreq=3361, maxDocs=44218)
                0.04550679 = queryNorm
              0.33530587 = fieldWeight in 1181, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.576596 = idf(docFreq=3361, maxDocs=44218)
                0.046875 = fieldNorm(doc=1181)
          0.03699325 = weight(_text_:22 in 1181) [ClassicSimilarity], result of:
            0.03699325 = score(doc=1181,freq=2.0), product of:
              0.15935703 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04550679 = queryNorm
              0.23214069 = fieldWeight in 1181, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=1181)
      0.5 = coord(1/2)
    
    Abstract
    The Nuovo soggettario, the official Italian subject indexing system edited by the National Central Library of Florence, is made up of interactive components, the core of which is a general thesaurus and some rules of a conventional syntax for subject string construction. The Nuovo soggettario Thesaurus is in compliance with ISO 25964: 2011-2013, IFLA LRM, and FAIR principle (findability, accessibility, interoperability, and reusability). Its open data are available in the Zthes, MARC21, and in SKOS formats and allow for interoperability with l library, archive, and museum databases. The Thesaurus's macrostructure is organized into four fundamental macro-categories, thirteen categories, and facets. The facets allow for the orderly development of hierarchies, thereby limiting polyhierarchies and promoting the grouping of homogenous concepts. This paper addresses the main features and peculiarities which have characterized the consistent development of this categorical structure and its effects on the syntactic sphere in a predominantly pre-coordinated usage context.
    Date
    26.11.2023 18:59:22
  3. Wu, Z.; Li, R.; Zhou, Z.; Guo, J.; Jiang, J.; Su, X.: ¬A user sensitive subject protection approach for book search service (2020) 0.03
    0.03149293 = product of:
      0.06298586 = sum of:
        0.06298586 = sum of:
          0.032158148 = weight(_text_:subject in 5617) [ClassicSimilarity], result of:
            0.032158148 = score(doc=5617,freq=2.0), product of:
              0.16275941 = queryWeight, product of:
                3.576596 = idf(docFreq=3361, maxDocs=44218)
                0.04550679 = queryNorm
              0.19758089 = fieldWeight in 5617, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.576596 = idf(docFreq=3361, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5617)
          0.03082771 = weight(_text_:22 in 5617) [ClassicSimilarity], result of:
            0.03082771 = score(doc=5617,freq=2.0), product of:
              0.15935703 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04550679 = queryNorm
              0.19345059 = fieldWeight in 5617, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5617)
      0.5 = coord(1/2)
    
    Date
    6. 1.2020 17:22:25
  4. Rae, A.R.; Mork, J.G.; Demner-Fushman, D.: ¬The National Library of Medicine indexer assignment dataset : a new large-scale dataset for reviewer assignment research (2023) 0.03
    0.03149293 = product of:
      0.06298586 = sum of:
        0.06298586 = sum of:
          0.032158148 = weight(_text_:subject in 885) [ClassicSimilarity], result of:
            0.032158148 = score(doc=885,freq=2.0), product of:
              0.16275941 = queryWeight, product of:
                3.576596 = idf(docFreq=3361, maxDocs=44218)
                0.04550679 = queryNorm
              0.19758089 = fieldWeight in 885, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.576596 = idf(docFreq=3361, maxDocs=44218)
                0.0390625 = fieldNorm(doc=885)
          0.03082771 = weight(_text_:22 in 885) [ClassicSimilarity], result of:
            0.03082771 = score(doc=885,freq=2.0), product of:
              0.15935703 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04550679 = queryNorm
              0.19345059 = fieldWeight in 885, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=885)
      0.5 = coord(1/2)
    
    Abstract
    MEDLINE is the National Library of Medicine's (NLM) journal citation database. It contains over 28 million references to biomedical and life science journal articles, and a key feature of the database is that all articles are indexed with NLM Medical Subject Headings (MeSH). The library employs a team of MeSH indexers, and in recent years they have been asked to index close to 1 million articles per year in order to keep MEDLINE up to date. An important part of the MEDLINE indexing process is the assignment of articles to indexers. High quality and timely indexing is only possible when articles are assigned to indexers with suitable expertise. This article introduces the NLM indexer assignment dataset: a large dataset of 4.2 million indexer article assignments for articles indexed between 2011 and 2019. The dataset is shown to be a valuable testbed for expert matching and assignment algorithms, and indexer article assignment is also found to be useful domain-adaptive pre-training for the closely related task of reviewer assignment.
    Date
    22. 1.2023 18:49:49
  5. Zhang, Y.; Liu, J.; Song, S.: ¬The design and evaluation of a nudge-based interface to facilitate consumers' evaluation of online health information credibility (2023) 0.03
    0.03149293 = product of:
      0.06298586 = sum of:
        0.06298586 = sum of:
          0.032158148 = weight(_text_:subject in 993) [ClassicSimilarity], result of:
            0.032158148 = score(doc=993,freq=2.0), product of:
              0.16275941 = queryWeight, product of:
                3.576596 = idf(docFreq=3361, maxDocs=44218)
                0.04550679 = queryNorm
              0.19758089 = fieldWeight in 993, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.576596 = idf(docFreq=3361, maxDocs=44218)
                0.0390625 = fieldNorm(doc=993)
          0.03082771 = weight(_text_:22 in 993) [ClassicSimilarity], result of:
            0.03082771 = score(doc=993,freq=2.0), product of:
              0.15935703 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04550679 = queryNorm
              0.19345059 = fieldWeight in 993, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=993)
      0.5 = coord(1/2)
    
    Abstract
    Evaluating the quality of online health information (OHI) is a major challenge facing consumers. We designed PageGraph, an interface that displays quality indicators and associated values for a webpage, based on credibility evaluation models, the nudge theory, and existing empirical research concerning professionals' and consumers' evaluation of OHI quality. A qualitative evaluation of the interface with 16 participants revealed that PageGraph rendered the information and presentation nudges as intended. It provided the participants with easier access to quality indicators, encouraged fresh angles to assess information credibility, provided an evaluation framework, and encouraged validation of initial judgments. We then conducted a quantitative evaluation of the interface involving 60 participants using a between-subject experimental design. The control group used a regular web browser and evaluated the credibility of 12 preselected webpages, whereas the experimental group evaluated the same webpages with the assistance of PageGraph. PageGraph did not significantly influence participants' evaluation results. The results may be attributed to the insufficiency of the saliency and structure of the nudges implemented and the webpage stimuli's lack of sensitivity to the intervention. Future directions for applying nudges to support OHI evaluation were discussed.
    Date
    22. 6.2023 18:18:34
  6. Barité, M.; Parentelli, V.; Rodríguez Casaballe, N.; Suárez, M.V.: Interdisciplinarity and postgraduate teaching of knowledge organization (KO) : elements for a necessary dialogue (2023) 0.03
    0.03149293 = product of:
      0.06298586 = sum of:
        0.06298586 = sum of:
          0.032158148 = weight(_text_:subject in 1125) [ClassicSimilarity], result of:
            0.032158148 = score(doc=1125,freq=2.0), product of:
              0.16275941 = queryWeight, product of:
                3.576596 = idf(docFreq=3361, maxDocs=44218)
                0.04550679 = queryNorm
              0.19758089 = fieldWeight in 1125, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.576596 = idf(docFreq=3361, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1125)
          0.03082771 = weight(_text_:22 in 1125) [ClassicSimilarity], result of:
            0.03082771 = score(doc=1125,freq=2.0), product of:
              0.15935703 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04550679 = queryNorm
              0.19345059 = fieldWeight in 1125, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1125)
      0.5 = coord(1/2)
    
    Abstract
    Interdisciplinarity implies the previous existence of disciplinary fields and not their dissolution. As a general objective, we propose to establish an initial approach to the emphasis given to interdisciplinarity in the teaching of KO, through the teaching staff responsible for postgraduate courses focused on -or related to the KO, in Ibero-American universities. For conducting the research, the framework and distribution of a survey addressed to teachers is proposed, based on four lines of action: 1. The way teachers manage the concept of interdisciplinarity. 2. The place that teachers give to interdisciplinarity in KO. 3. Assessment of interdisciplinary content that teachers incorporate into their postgraduate courses. 4. Set of teaching strategies and resources used by teachers to include interdisciplinarity in the teaching of KO. The study analyzed 22 responses. Preliminary results show that KO teachers recognize the influence of other disciplines in concepts, theories, methods, and applications, but no consensus has been reached regarding which disciplines and authors are the ones who build interdisciplinary bridges. Among other conclusions, the study strongly suggests that environmental and social tensions are reflected in subject representation, especially in the construction of friendly knowl­edge organization systems with interdisciplinary visions, and in the expressions through which information is sought.
  7. Marques Redigolo, F.; Lopes Fujita, M.S.; Gil-Leiva, I.: Guidelines for subject analysis in subject cataloging (2022) 0.03
    0.028942334 = product of:
      0.057884667 = sum of:
        0.057884667 = product of:
          0.115769334 = sum of:
            0.115769334 = weight(_text_:subject in 736) [ClassicSimilarity], result of:
              0.115769334 = score(doc=736,freq=18.0), product of:
                0.16275941 = queryWeight, product of:
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.04550679 = queryNorm
                0.7112912 = fieldWeight in 736, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.046875 = fieldNorm(doc=736)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The representation of information in subject cataloging as a result of subject analysis will depend on the cataloger's prior knowledge, influenced by subjectivity. The subject analysis in cataloging is a central theme of this investigation with the aim to elaborate guidelines for subject analysis in cataloging. For this purpose, how books are cataloged in university libraries has been verified. The Individual Verbal Protocol was applied with catalogers from Brazilian and Spanish University Libraries. Directions for the elements and variables of the subject analysis and procedures for good development were obtained to constitute the Guidelines of Subject Analysis in Cataloging. It is concluded that the guidelines formed by four sections are indicated for incorporation in subject cataloging procedure manuals for the purpose of improving the levels of representation and information retrieval results.
  8. Wu, M.; Liu, Y.-H.; Brownlee, R.; Zhang, X.: Evaluating utility and automatic classification of subject metadata from Research Data Australia (2021) 0.03
    0.027287092 = product of:
      0.054574184 = sum of:
        0.054574184 = product of:
          0.10914837 = sum of:
            0.10914837 = weight(_text_:subject in 453) [ClassicSimilarity], result of:
              0.10914837 = score(doc=453,freq=16.0), product of:
                0.16275941 = queryWeight, product of:
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.04550679 = queryNorm
                0.67061174 = fieldWeight in 453, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.046875 = fieldNorm(doc=453)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    In this paper, we present a case study of how well subject metadata (comprising headings from an international classification scheme) has been deployed in a national data catalogue, and how often data seekers use subject metadata when searching for data. Through an analysis of user search behaviour as recorded in search logs, we find evidence that users utilise the subject metadata for data discovery. Since approximately half of the records ingested by the catalogue did not include subject metadata at the time of harvest, we experimented with automatic subject classification approaches in order to enrich these records and to provide additional support for user search and data discovery. Our results show that automatic methods work well for well represented categories of subject metadata, and these categories tend to have features that can distinguish themselves from the other categories. Our findings raise implications for data catalogue providers; they should invest more effort to enhance the quality of data records by providing an adequate description of these records for under-represented subject categories.
  9. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.03
    0.027103817 = product of:
      0.054207634 = sum of:
        0.054207634 = product of:
          0.21683054 = sum of:
            0.21683054 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
              0.21683054 = score(doc=862,freq=2.0), product of:
                0.38580707 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.04550679 = queryNorm
                0.56201804 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=862)
          0.25 = coord(1/4)
      0.5 = coord(1/2)
    
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
  10. Hutchinson, J.; Nakatomi, J.: Improving subject description of an LGBTQ+ collection (2024) 0.03
    0.025726518 = product of:
      0.051453035 = sum of:
        0.051453035 = product of:
          0.10290607 = sum of:
            0.10290607 = weight(_text_:subject in 1157) [ClassicSimilarity], result of:
              0.10290607 = score(doc=1157,freq=8.0), product of:
                0.16275941 = queryWeight, product of:
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.04550679 = queryNorm
                0.63225883 = fieldWeight in 1157, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1157)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This article summarizes the work done as part of a project to improve subject description of an LGBTQ + collection in the ONE Archives, part of the University of Southern California (USC) Libraries. The project involved adding local subject headings to augment existing Library of Congress Subject Headings. The article describes the steps that the project team took, along with the methods that were rejected. The paper discusses reasons why the team chose their course of action.
  11. Asula, M.; Makke, J.; Freienthal, L.; Kuulmets, H.-A.; Sirel, R.: Kratt: developing an automatic subject indexing tool for the National Library of Estonia : how to transfer metadata information among work cluster members (2021) 0.03
    0.025524741 = product of:
      0.051049482 = sum of:
        0.051049482 = product of:
          0.102098964 = sum of:
            0.102098964 = weight(_text_:subject in 723) [ClassicSimilarity], result of:
              0.102098964 = score(doc=723,freq=14.0), product of:
                0.16275941 = queryWeight, product of:
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.04550679 = queryNorm
                0.6272999 = fieldWeight in 723, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.046875 = fieldNorm(doc=723)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Manual subject indexing in libraries is a time-consuming and costly process and the quality of the assigned subjects is affected by the cataloger's knowledge on the specific topics contained in the book. Trying to solve these issues, we exploited the opportunities arising from artificial intelligence to develop Kratt: a prototype of an automatic subject indexing tool. Kratt is able to subject index a book independent of its extent and genre with a set of keywords present in the Estonian Subject Thesaurus. It takes Kratt approximately one minute to subject index a book, outperforming humans 10-15 times. Although the resulting keywords were not considered satisfactory by the catalogers, the ratings of a small sample of regular library users showed more promise. We also argue that the results can be enhanced by including a bigger corpus for training the model and applying more careful preprocessing techniques.
    Footnote
    Teil eines Themenheftes: Artificial intelligence (AI) and automated processes for subject sccess
  12. Golub, K.; Tyrkkö, J.; Hansson, J.; Ahlström, I.: Subject indexing in humanities : a comparison between a local university repository and an international bibliographic service (2020) 0.02
    0.022739245 = product of:
      0.04547849 = sum of:
        0.04547849 = product of:
          0.09095698 = sum of:
            0.09095698 = weight(_text_:subject in 5982) [ClassicSimilarity], result of:
              0.09095698 = score(doc=5982,freq=16.0), product of:
                0.16275941 = queryWeight, product of:
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.04550679 = queryNorm
                0.55884314 = fieldWeight in 5982, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5982)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    As the humanities develop in the realm of increasingly more pronounced digital scholarship, it is important to provide quality subject access to a vast range of heterogeneous information objects in digital services. The study aims to paint a representative picture of the current state of affairs of the use of subject index terms in humanities journal articles with particular reference to the well-established subject access needs of humanities researchers, with the purpose of identifying which improvements are needed in this context. Design/methodology/approach The comparison of subject metadata on a sample of 649 peer-reviewed journal articles from across the humanities is conducted in a university repository, against Scopus, the former reflecting local and national policies and the latter being the most comprehensive international abstract and citation database of research output. Findings The study shows that established bibliographic objectives to ensure subject access for humanities journal articles are not supported in either the world's largest commercial abstract and citation database Scopus or the local repository of a public university in Sweden. The indexing policies in the two services do not seem to address the needs of humanities scholars for highly granular subject index terms with appropriate facets; no controlled vocabularies for any humanities discipline are used whatsoever. Originality/value In all, not much has changed since 1990s when indexing for the humanities was shown to lag behind the sciences. The community of researchers and information professionals, today working together on digital humanities projects, as well as interdisciplinary research teams, should demand that their subject access needs be fulfilled, especially in commercial services like Scopus and discovery services.
  13. Dietz, K.: en.wikipedia.org > 6 Mio. Artikel (2020) 0.02
    0.022586515 = product of:
      0.04517303 = sum of:
        0.04517303 = product of:
          0.18069212 = sum of:
            0.18069212 = weight(_text_:3a in 5669) [ClassicSimilarity], result of:
              0.18069212 = score(doc=5669,freq=2.0), product of:
                0.38580707 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.04550679 = queryNorm
                0.46834838 = fieldWeight in 5669, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5669)
          0.25 = coord(1/4)
      0.5 = coord(1/2)
    
    Content
    "Die Englischsprachige Wikipedia verfügt jetzt über mehr als 6 Millionen Artikel. An zweiter Stelle kommt die deutschsprachige Wikipedia mit 2.3 Millionen Artikeln, an dritter Stelle steht die französischsprachige Wikipedia mit 2.1 Millionen Artikeln (via Researchbuzz: Firehose <https://rbfirehose.com/2020/01/24/techcrunch-wikipedia-now-has-more-than-6-million-articles-in-english/> und Techcrunch <https://techcrunch.com/2020/01/23/wikipedia-english-six-million-articles/?utm_source=feedburner&utm_medium=feed&utm_campaign=Feed%3A+Techcrunch+%28TechCrunch%29&guccounter=1&guce_referrer=aHR0cHM6Ly9yYmZpcmVob3NlLmNvbS8yMDIwLzAxLzI0L3RlY2hjcnVuY2gtd2lraXBlZGlhLW5vdy1oYXMtbW9yZS10aGFuLTYtbWlsbGlvbi1hcnRpY2xlcy1pbi1lbmdsaXNoLw&guce_referrer_sig=AQAAAK0zHfjdDZ_spFZBF_z-zDjtL5iWvuKDumFTzm4HvQzkUfE2pLXQzGS6FGB_y-VISdMEsUSvkNsg2U_NWQ4lwWSvOo3jvXo1I3GtgHpP8exukVxYAnn5mJspqX50VHIWFADHhs5AerkRn3hMRtf_R3F1qmEbo8EROZXp328HMC-o>). 250120 via digithek ch = #fineBlog s.a.: Angesichts der Veröffentlichung des 6-millionsten Artikels vergangene Woche in der englischsprachigen Wikipedia hat die Community-Zeitungsseite "Wikipedia Signpost" ein Moratorium bei der Veröffentlichung von Unternehmensartikeln gefordert. Das sei kein Vorwurf gegen die Wikimedia Foundation, aber die derzeitigen Maßnahmen, um die Enzyklopädie gegen missbräuchliches undeklariertes Paid Editing zu schützen, funktionierten ganz klar nicht. *"Da die ehrenamtlichen Autoren derzeit von Werbung in Gestalt von Wikipedia-Artikeln überwältigt werden, und da die WMF nicht in der Lage zu sein scheint, dem irgendetwas entgegenzusetzen, wäre der einzige gangbare Weg für die Autoren, fürs erste die Neuanlage von Artikeln über Unternehmen zu untersagen"*, schreibt der Benutzer Smallbones in seinem Editorial <https://en.wikipedia.org/wiki/Wikipedia:Wikipedia_Signpost/2020-01-27/From_the_editor> zur heutigen Ausgabe."
  14. Gabler, S.: Vergabe von DDC-Sachgruppen mittels eines Schlagwort-Thesaurus (2021) 0.02
    0.022586515 = product of:
      0.04517303 = sum of:
        0.04517303 = product of:
          0.18069212 = sum of:
            0.18069212 = weight(_text_:3a in 1000) [ClassicSimilarity], result of:
              0.18069212 = score(doc=1000,freq=2.0), product of:
                0.38580707 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.04550679 = queryNorm
                0.46834838 = fieldWeight in 1000, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1000)
          0.25 = coord(1/4)
      0.5 = coord(1/2)
    
    Content
    Master thesis Master of Science (Library and Information Studies) (MSc), Universität Wien. Advisor: Christoph Steiner. Vgl.: https://www.researchgate.net/publication/371680244_Vergabe_von_DDC-Sachgruppen_mittels_eines_Schlagwort-Thesaurus. DOI: 10.25365/thesis.70030. Vgl. dazu die Präsentation unter: https://www.google.com/url?sa=i&rct=j&q=&esrc=s&source=web&cd=&ved=0CAIQw7AJahcKEwjwoZzzytz_AhUAAAAAHQAAAAAQAg&url=https%3A%2F%2Fwiki.dnb.de%2Fdownload%2Fattachments%2F252121510%2FDA3%2520Workshop-Gabler.pdf%3Fversion%3D1%26modificationDate%3D1671093170000%26api%3Dv2&psig=AOvVaw0szwENK1or3HevgvIDOfjx&ust=1687719410889597&opi=89978449.
  15. Golub, K.: Automated subject indexing : an overview (2021) 0.02
    0.022510704 = product of:
      0.045021407 = sum of:
        0.045021407 = product of:
          0.090042815 = sum of:
            0.090042815 = weight(_text_:subject in 718) [ClassicSimilarity], result of:
              0.090042815 = score(doc=718,freq=8.0), product of:
                0.16275941 = queryWeight, product of:
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.04550679 = queryNorm
                0.5532265 = fieldWeight in 718, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=718)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    In the face of the ever-increasing document volume, libraries around the globe are more and more exploring (semi-) automated approaches to subject indexing. This helps sustain bibliographic objectives, enrich metadata, and establish more connections across documents from various collections, effectively leading to improved information retrieval and access. However, generally accepted automated approaches that are functional in operative systems are lacking. This article aims to provide an overview of basic principles used for automated subject indexing, major approaches in relation to their possible application in actual library systems, existing working examples, as well as related challenges calling for further research.
    Footnote
    Teil eines Themenheftes: Artificial intelligence (AI) and automated processes for subject sccess
  16. Chou, C.; Chu, T.: ¬An analysis of BERT (NLP) for assisted subject indexing for Project Gutenberg (2022) 0.02
    0.022510704 = product of:
      0.045021407 = sum of:
        0.045021407 = product of:
          0.090042815 = sum of:
            0.090042815 = weight(_text_:subject in 1139) [ClassicSimilarity], result of:
              0.090042815 = score(doc=1139,freq=8.0), product of:
                0.16275941 = queryWeight, product of:
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.04550679 = queryNorm
                0.5532265 = fieldWeight in 1139, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1139)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    In light of AI (Artificial Intelligence) and NLP (Natural language processing) technologies, this article examines the feasibility of using AI/NLP models to enhance the subject indexing of digital resources. While BERT (Bidirectional Encoder Representations from Transformers) models are widely used in scholarly communities, the authors assess whether BERT models can be used in machine-assisted indexing in the Project Gutenberg collection, through suggesting Library of Congress subject headings filtered by certain Library of Congress Classification subclass labels. The findings of this study are informative for further research on BERT models to assist with automatic subject indexing for digital library collections.
  17. Cooey, N.; Phillips, A.: Library of Congress Subject Headings : a post-coordinated future (2023) 0.02
    0.022510704 = product of:
      0.045021407 = sum of:
        0.045021407 = product of:
          0.090042815 = sum of:
            0.090042815 = weight(_text_:subject in 1163) [ClassicSimilarity], result of:
              0.090042815 = score(doc=1163,freq=8.0), product of:
                0.16275941 = queryWeight, product of:
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.04550679 = queryNorm
                0.5532265 = fieldWeight in 1163, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1163)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This paper is the result of a request from Library of Congress leadership to assess pre-coordinated versus post-coordinated subject cataloging. It argues that the disadvantages of pre-coordinated subject strings are perennial and continue to hinder progress, while the advantages of post-coordinated subject cataloging have expanded, resulting in new opportunities to serve the needs of catalogers and end users alike. The consequences of retaining pre-coordinated headings will have long-term impacts that heavily out-weigh the short-term challenges of transitioning to new cataloging practices. By implementing post-coordinated, faceted vocabularies, the Library of Congress will be investing in the future of libraries.
  18. Zimmerman, N.: User study: implementation of OCLC FAST subject headings in the Lafayette digital repository (2023) 0.02
    0.022510704 = product of:
      0.045021407 = sum of:
        0.045021407 = product of:
          0.090042815 = sum of:
            0.090042815 = weight(_text_:subject in 1176) [ClassicSimilarity], result of:
              0.090042815 = score(doc=1176,freq=8.0), product of:
                0.16275941 = queryWeight, product of:
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.04550679 = queryNorm
                0.5532265 = fieldWeight in 1176, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1176)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Digital repository migrations present a periodic opportunity to assess metadata quality and to perform strategic enhancements. Lafayette College Libraries implemented OCLC FAST (Faceted Application of Subject Terminology) for its digital image collections as part of a migration from multiple repositories to a single one built on the Samvera Hyrax open-source framework. Application of FAST has normalized subject headings across dissimilar collections in a way that tremendously improves descriptive consistency for staff and discoverability for end users. However, the process of applying FAST headings was complicated by several features of in-scope metadata as well as gaps in available controlled subject authorities.
  19. ¬Der Student aus dem Computer (2023) 0.02
    0.021579396 = product of:
      0.043158792 = sum of:
        0.043158792 = product of:
          0.086317584 = sum of:
            0.086317584 = weight(_text_:22 in 1079) [ClassicSimilarity], result of:
              0.086317584 = score(doc=1079,freq=2.0), product of:
                0.15935703 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04550679 = queryNorm
                0.5416616 = fieldWeight in 1079, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1079)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    27. 1.2023 16:22:55
  20. Ahmed, M.: Automatic indexing for agriculture : designing a framework by deploying Agrovoc, Agris and Annif (2023) 0.02
    0.019692764 = product of:
      0.039385527 = sum of:
        0.039385527 = product of:
          0.078771055 = sum of:
            0.078771055 = weight(_text_:subject in 1024) [ClassicSimilarity], result of:
              0.078771055 = score(doc=1024,freq=12.0), product of:
                0.16275941 = queryWeight, product of:
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.04550679 = queryNorm
                0.48397237 = fieldWeight in 1024, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  3.576596 = idf(docFreq=3361, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1024)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    There are several ways to employ machine learning for automating subject indexing. One popular strategy is to utilize a supervised learning algorithm to train a model on a set of documents that have been manually indexed by subject matter using a standard vocabulary. The resulting model can then predict the subject of new and previously unseen documents by identifying patterns learned from the training data. To do this, the first step is to gather a large dataset of documents and manually assign each document a set of subject keywords/descriptors from a controlled vocabulary (e.g., from Agrovoc). Next, the dataset (obtained from Agris) can be divided into - i) a training dataset, and ii) a test dataset. The training dataset is used to train the model, while the test dataset is used to evaluate the model's performance. Machine learning can be a powerful tool for automating the process of subject indexing. This research is an attempt to apply Annif (http://annif. org/), an open-source AI/ML framework, to autogenerate subject keywords/descriptors for documentary resources in the domain of agriculture. The training dataset is obtained from Agris, which applies the Agrovoc thesaurus as a vocabulary tool (https://www.fao.org/agris/download).

Languages

  • e 158
  • d 31

Types

  • a 178
  • el 28
  • p 5
  • m 3
  • x 1
  • More… Less…