Search (125 results, page 1 of 7)

  • × year_i:[2020 TO 2030}
  1. Candela, G.: ¬An automatic data quality approach to assess semantic data from cultural heritage institutions (2023) 0.15
    0.1484918 = product of:
      0.2969836 = sum of:
        0.2376742 = weight(_text_:assess in 997) [ClassicSimilarity], result of:
          0.2376742 = score(doc=997,freq=4.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.64474034 = fieldWeight in 997, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0546875 = fieldNorm(doc=997)
        0.059309397 = weight(_text_:22 in 997) [ClassicSimilarity], result of:
          0.059309397 = score(doc=997,freq=2.0), product of:
            0.21899058 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.062536046 = queryNorm
            0.2708308 = fieldWeight in 997, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0546875 = fieldNorm(doc=997)
      0.5 = coord(2/4)
    
    Abstract
    In recent years, cultural heritage institutions have been exploring the benefits of applying Linked Open Data to their catalogs and digital materials. Innovative and creative methods have emerged to publish and reuse digital contents to promote computational access, such as the concepts of Labs and Collections as Data. Data quality has become a requirement for researchers and training methods based on artificial intelligence and machine learning. This article explores how the quality of Linked Open Data made available by cultural heritage institutions can be automatically assessed. The results obtained can be useful for other institutions who wish to publish and assess their collections.
    Date
    22. 6.2023 18:23:31
  2. Zhang, Y.; Liu, J.; Song, S.: ¬The design and evaluation of a nudge-based interface to facilitate consumers' evaluation of online health information credibility (2023) 0.08
    0.08120373 = product of:
      0.16240746 = sum of:
        0.120043606 = weight(_text_:assess in 993) [ClassicSimilarity], result of:
          0.120043606 = score(doc=993,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.32564306 = fieldWeight in 993, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=993)
        0.042363856 = weight(_text_:22 in 993) [ClassicSimilarity], result of:
          0.042363856 = score(doc=993,freq=2.0), product of:
            0.21899058 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.062536046 = queryNorm
            0.19345059 = fieldWeight in 993, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0390625 = fieldNorm(doc=993)
      0.5 = coord(2/4)
    
    Abstract
    Evaluating the quality of online health information (OHI) is a major challenge facing consumers. We designed PageGraph, an interface that displays quality indicators and associated values for a webpage, based on credibility evaluation models, the nudge theory, and existing empirical research concerning professionals' and consumers' evaluation of OHI quality. A qualitative evaluation of the interface with 16 participants revealed that PageGraph rendered the information and presentation nudges as intended. It provided the participants with easier access to quality indicators, encouraged fresh angles to assess information credibility, provided an evaluation framework, and encouraged validation of initial judgments. We then conducted a quantitative evaluation of the interface involving 60 participants using a between-subject experimental design. The control group used a regular web browser and evaluated the credibility of 12 preselected webpages, whereas the experimental group evaluated the same webpages with the assistance of PageGraph. PageGraph did not significantly influence participants' evaluation results. The results may be attributed to the insufficiency of the saliency and structure of the nudges implemented and the webpage stimuli's lack of sensitivity to the intervention. Future directions for applying nudges to support OHI evaluation were discussed.
    Date
    22. 6.2023 18:18:34
  3. Wang, J.; Halffman, W.; Zhang, Y.H.: Sorting out journals : the proliferation of journal lists in China (2023) 0.08
    0.08120373 = product of:
      0.16240746 = sum of:
        0.120043606 = weight(_text_:assess in 1055) [ClassicSimilarity], result of:
          0.120043606 = score(doc=1055,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.32564306 = fieldWeight in 1055, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1055)
        0.042363856 = weight(_text_:22 in 1055) [ClassicSimilarity], result of:
          0.042363856 = score(doc=1055,freq=2.0), product of:
            0.21899058 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.062536046 = queryNorm
            0.19345059 = fieldWeight in 1055, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1055)
      0.5 = coord(2/4)
    
    Abstract
    Journal lists are instruments to categorize, compare, and assess research and scholarly publications. Our study investigates the remarkable proliferation of such journal lists in China, analyses their underlying values, quality criteria and ranking principles, and specifies how concerns specific to the Chinese research policy and publishing system inform these lists. Discouraged lists of "bad journals" reflect concerns over inferior research publications, but also the involved drain on public resources. Endorsed lists of "good journals" are based on criteria valued in research policy, reflecting the distinctive administrative logic of state-led Chinese research and publishing policy, ascribing worth to scientific journals for its specific national and institutional needs. In this regard, the criteria used for journal list construction are contextual and reflect the challenges of public resource allocation in a market-led publication system. Chinese journal lists therefore reflect research policy changes, such as a shift away from output-dominated research evaluation, the specific concerns about research misconduct, and balancing national research needs against international standards, resulting in distinctly Chinese quality criteria. However, contrasting concerns and inaccuracies lead to contradictions in the "qualify" and "disqualify" binary logic and demonstrate inherent tensions and limitations in journal lists as policy tools.
    Date
    22. 9.2023 16:39:23
  4. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.07
    0.074492864 = product of:
      0.29797146 = sum of:
        0.29797146 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
          0.29797146 = score(doc=862,freq=2.0), product of:
            0.5301813 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.062536046 = queryNorm
            0.56201804 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
      0.25 = coord(1/4)
    
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
  5. Dietz, K.: en.wikipedia.org > 6 Mio. Artikel (2020) 0.06
    0.06207739 = product of:
      0.24830955 = sum of:
        0.24830955 = weight(_text_:3a in 5669) [ClassicSimilarity], result of:
          0.24830955 = score(doc=5669,freq=2.0), product of:
            0.5301813 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.062536046 = queryNorm
            0.46834838 = fieldWeight in 5669, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5669)
      0.25 = coord(1/4)
    
    Content
    "Die Englischsprachige Wikipedia verfügt jetzt über mehr als 6 Millionen Artikel. An zweiter Stelle kommt die deutschsprachige Wikipedia mit 2.3 Millionen Artikeln, an dritter Stelle steht die französischsprachige Wikipedia mit 2.1 Millionen Artikeln (via Researchbuzz: Firehose <https://rbfirehose.com/2020/01/24/techcrunch-wikipedia-now-has-more-than-6-million-articles-in-english/> und Techcrunch <https://techcrunch.com/2020/01/23/wikipedia-english-six-million-articles/?utm_source=feedburner&utm_medium=feed&utm_campaign=Feed%3A+Techcrunch+%28TechCrunch%29&guccounter=1&guce_referrer=aHR0cHM6Ly9yYmZpcmVob3NlLmNvbS8yMDIwLzAxLzI0L3RlY2hjcnVuY2gtd2lraXBlZGlhLW5vdy1oYXMtbW9yZS10aGFuLTYtbWlsbGlvbi1hcnRpY2xlcy1pbi1lbmdsaXNoLw&guce_referrer_sig=AQAAAK0zHfjdDZ_spFZBF_z-zDjtL5iWvuKDumFTzm4HvQzkUfE2pLXQzGS6FGB_y-VISdMEsUSvkNsg2U_NWQ4lwWSvOo3jvXo1I3GtgHpP8exukVxYAnn5mJspqX50VHIWFADHhs5AerkRn3hMRtf_R3F1qmEbo8EROZXp328HMC-o>). 250120 via digithek ch = #fineBlog s.a.: Angesichts der Veröffentlichung des 6-millionsten Artikels vergangene Woche in der englischsprachigen Wikipedia hat die Community-Zeitungsseite "Wikipedia Signpost" ein Moratorium bei der Veröffentlichung von Unternehmensartikeln gefordert. Das sei kein Vorwurf gegen die Wikimedia Foundation, aber die derzeitigen Maßnahmen, um die Enzyklopädie gegen missbräuchliches undeklariertes Paid Editing zu schützen, funktionierten ganz klar nicht. *"Da die ehrenamtlichen Autoren derzeit von Werbung in Gestalt von Wikipedia-Artikeln überwältigt werden, und da die WMF nicht in der Lage zu sein scheint, dem irgendetwas entgegenzusetzen, wäre der einzige gangbare Weg für die Autoren, fürs erste die Neuanlage von Artikeln über Unternehmen zu untersagen"*, schreibt der Benutzer Smallbones in seinem Editorial <https://en.wikipedia.org/wiki/Wikipedia:Wikipedia_Signpost/2020-01-27/From_the_editor> zur heutigen Ausgabe."
  6. Gabler, S.: Vergabe von DDC-Sachgruppen mittels eines Schlagwort-Thesaurus (2021) 0.06
    0.06207739 = product of:
      0.24830955 = sum of:
        0.24830955 = weight(_text_:3a in 1000) [ClassicSimilarity], result of:
          0.24830955 = score(doc=1000,freq=2.0), product of:
            0.5301813 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.062536046 = queryNorm
            0.46834838 = fieldWeight in 1000, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1000)
      0.25 = coord(1/4)
    
    Content
    Master thesis Master of Science (Library and Information Studies) (MSc), Universität Wien. Advisor: Christoph Steiner. Vgl.: https://www.researchgate.net/publication/371680244_Vergabe_von_DDC-Sachgruppen_mittels_eines_Schlagwort-Thesaurus. DOI: 10.25365/thesis.70030. Vgl. dazu die Präsentation unter: https://www.google.com/url?sa=i&rct=j&q=&esrc=s&source=web&cd=&ved=0CAIQw7AJahcKEwjwoZzzytz_AhUAAAAAHQAAAAAQAg&url=https%3A%2F%2Fwiki.dnb.de%2Fdownload%2Fattachments%2F252121510%2FDA3%2520Workshop-Gabler.pdf%3Fversion%3D1%26modificationDate%3D1671093170000%26api%3Dv2&psig=AOvVaw0szwENK1or3HevgvIDOfjx&ust=1687719410889597&opi=89978449.
  7. Butlin, P.; Long, R.; Elmoznino, E.; Bengio, Y.; Birch, J.; Constant, A.; Deane, G.; Fleming, S.M.; Frith, C.; Ji, X.; Kanai, R.; Klein, C.; Lindsay, G.; Michel, M.; Mudrik, L.; Peters, M.A.K.; Schwitzgebel, E.; Simon, J.; VanRullen, R.: Consciousness in artificial intelligence : insights from the science of consciousness (2023) 0.05
    0.050930187 = product of:
      0.20372075 = sum of:
        0.20372075 = weight(_text_:assess in 1214) [ClassicSimilarity], result of:
          0.20372075 = score(doc=1214,freq=4.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.5526346 = fieldWeight in 1214, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.046875 = fieldNorm(doc=1214)
      0.25 = coord(1/4)
    
    Abstract
    Whether current or near-term AI systems could be conscious is a topic of scientific interest and increasing public concern. This report argues for, and exemplifies, a rigorous and empirically grounded approach to AI consciousness: assessing existing AI systems in detail, in light of our best-supported neuroscientific theories of consciousness. We survey several prominent scientific theories of consciousness, including recurrent processing theory, global workspace theory, higher-order theories, predictive processing, and attention schema theory. From these theories we derive "indicator properties" of consciousness, elucidated in computational terms that allow us to assess AI systems for these properties. We use these indicator properties to assess several recent AI systems, and we discuss how future systems might implement them. Our analysis suggests that no current AI systems are conscious, but also suggests that there are no obvious technical barriers to building AI systems which satisfy these indicators.
  8. Wang, P.; Li, X.: Assessing the quality of information on Wikipedia : a deep-learning approach (2020) 0.04
    0.04244182 = product of:
      0.16976728 = sum of:
        0.16976728 = weight(_text_:assess in 5505) [ClassicSimilarity], result of:
          0.16976728 = score(doc=5505,freq=4.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.4605288 = fieldWeight in 5505, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5505)
      0.25 = coord(1/4)
    
    Abstract
    Currently, web document repositories have been collaboratively created and edited. One of these repositories, Wikipedia, is facing an important problem: assessing the quality of Wikipedia. Existing approaches exploit techniques such as statistical models or machine leaning algorithms to assess Wikipedia article quality. However, existing models do not provide satisfactory results. Furthermore, these models fail to adopt a comprehensive feature framework. In this article, we conduct an extensive survey of previous studies and summarize a comprehensive feature framework, including text statistics, writing style, readability, article structure, network, and editing history. Selected state-of-the-art deep-learning models, including the convolutional neural network (CNN), deep neural network (DNN), long short-term memory (LSTMs) network, CNN-LSTMs, bidirectional LSTMs, and stacked LSTMs, are applied to assess the quality of Wikipedia. A detailed comparison of deep-learning models is conducted with regard to different aspects: classification performance and training performance. We include an importance analysis of different features and feature sets to determine which features or feature sets are most effective in distinguishing Wikipedia article quality. This extensive experiment validates the effectiveness of the proposed model.
  9. Trace, C.B.; Zhang, Y.; Yi, S.; Williams-Brown, M.Y.: Information practices around genetic testing for ovarian cancer patients (2023) 0.04
    0.04244182 = product of:
      0.16976728 = sum of:
        0.16976728 = weight(_text_:assess in 1071) [ClassicSimilarity], result of:
          0.16976728 = score(doc=1071,freq=4.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.4605288 = fieldWeight in 1071, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1071)
      0.25 = coord(1/4)
    
    Abstract
    Knowledge of ovarian cancer patients' information practices around cancer genetic testing (GT) is needed to inform interventions that promote patient access to GT-related information. We interviewed 21 ovarian cancer patients and survivors who had GT as part of the treatment process and analyzed the transcripts using the qualitative content analysis method. We found that patients' information practices, manifested in their information-seeking mode, information sources utilized, information assessment, and information use, showed three distinct styles: passive, semi-active, and active. Patients with the passive style primarily received information from clinical sources, encountered information, or delegated information-seeking to family members; they were not inclined to assess information themselves and seldom used it to learn or influence others. Women with semi-active and active styles adopted more active information-seeking modes to approach information, utilized information sources beyond clinical settings, attempted to assess the information found, and actively used it to learn, educate others, or advocate GT to family and friends. Guided by the social ecological model, we found multiple levels of influences, including personal, interpersonal, organizational, community, and societal, acting as motivators or barriers to patients' information practice. Based on these findings, we discussed strategies to promote patient access to GT-related information.
  10. Hudon, M.: ¬The status of knowledge organization in library and information science master's programs (2021) 0.04
    0.04201526 = product of:
      0.16806103 = sum of:
        0.16806103 = weight(_text_:assess in 697) [ClassicSimilarity], result of:
          0.16806103 = score(doc=697,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.45590025 = fieldWeight in 697, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0546875 = fieldNorm(doc=697)
      0.25 = coord(1/4)
    
    Abstract
    The content of master's programs accredited by the American Library Association was examined to assess the status of knowledge organization (KO) as a subject in current training. Data collected show that KO remains very visible in a majority of programs, mainly in the form of required and electives courses focusing on descriptive cataloging, classification, and metadata. Observed tendencies include, however, the recent elimination of the required KO course in several programs, the reality that one third of KO electives listed in course catalogs have not been scheduled in the past three years, and the fact that two-thirds of those teaching KO specialize in other areas of information science.
  11. Chou, C.; Chu, T.: ¬An analysis of BERT (NLP) for assisted subject indexing for Project Gutenberg (2022) 0.04
    0.04201526 = product of:
      0.16806103 = sum of:
        0.16806103 = weight(_text_:assess in 1139) [ClassicSimilarity], result of:
          0.16806103 = score(doc=1139,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.45590025 = fieldWeight in 1139, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1139)
      0.25 = coord(1/4)
    
    Abstract
    In light of AI (Artificial Intelligence) and NLP (Natural language processing) technologies, this article examines the feasibility of using AI/NLP models to enhance the subject indexing of digital resources. While BERT (Bidirectional Encoder Representations from Transformers) models are widely used in scholarly communities, the authors assess whether BERT models can be used in machine-assisted indexing in the Project Gutenberg collection, through suggesting Library of Congress subject headings filtered by certain Library of Congress Classification subclass labels. The findings of this study are informative for further research on BERT models to assist with automatic subject indexing for digital library collections.
  12. Bullard, J.; Watson, B.; Purdome, C.: Misrepresentation in the surrogate : author critiques of "Indians of North America" subject headings (2022) 0.04
    0.04201526 = product of:
      0.16806103 = sum of:
        0.16806103 = weight(_text_:assess in 1143) [ClassicSimilarity], result of:
          0.16806103 = score(doc=1143,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.45590025 = fieldWeight in 1143, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1143)
      0.25 = coord(1/4)
    
    Abstract
    The surrogate record for a book in the library catalog contains subject headings applied on the basis of literary warrant. To assess the extent to which terms like "Indians of North America" are accurate to the content of the items with that label, we invited the items' creators to critique their surrogate records. In interviews with 38 creators we found consensus against the term "Indians of North America" and identified a periphery of related terms that misrepresent the content of the work, are out of alignment with their scholarly communities, and reproduce settler colonial biases in our library systems.
  13. Cooey, N.; Phillips, A.: Library of Congress Subject Headings : a post-coordinated future (2023) 0.04
    0.04201526 = product of:
      0.16806103 = sum of:
        0.16806103 = weight(_text_:assess in 1163) [ClassicSimilarity], result of:
          0.16806103 = score(doc=1163,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.45590025 = fieldWeight in 1163, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1163)
      0.25 = coord(1/4)
    
    Abstract
    This paper is the result of a request from Library of Congress leadership to assess pre-coordinated versus post-coordinated subject cataloging. It argues that the disadvantages of pre-coordinated subject strings are perennial and continue to hinder progress, while the advantages of post-coordinated subject cataloging have expanded, resulting in new opportunities to serve the needs of catalogers and end users alike. The consequences of retaining pre-coordinated headings will have long-term impacts that heavily out-weigh the short-term challenges of transitioning to new cataloging practices. By implementing post-coordinated, faceted vocabularies, the Library of Congress will be investing in the future of libraries.
  14. Zimmerman, N.: User study: implementation of OCLC FAST subject headings in the Lafayette digital repository (2023) 0.04
    0.04201526 = product of:
      0.16806103 = sum of:
        0.16806103 = weight(_text_:assess in 1176) [ClassicSimilarity], result of:
          0.16806103 = score(doc=1176,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.45590025 = fieldWeight in 1176, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1176)
      0.25 = coord(1/4)
    
    Abstract
    Digital repository migrations present a periodic opportunity to assess metadata quality and to perform strategic enhancements. Lafayette College Libraries implemented OCLC FAST (Faceted Application of Subject Terminology) for its digital image collections as part of a migration from multiple repositories to a single one built on the Samvera Hyrax open-source framework. Application of FAST has normalized subject headings across dissimilar collections in a way that tremendously improves descriptive consistency for staff and discoverability for end users. However, the process of applying FAST headings was complicated by several features of in-scope metadata as well as gaps in available controlled subject authorities.
  15. Menkov, V.; Ginsparg, P.; Kantor, P.B.: Recommendations and privacy in the arXiv system : a simulation experiment using historical data (2020) 0.04
    0.036013078 = product of:
      0.14405231 = sum of:
        0.14405231 = weight(_text_:assess in 5671) [ClassicSimilarity], result of:
          0.14405231 = score(doc=5671,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.39077166 = fieldWeight in 5671, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.046875 = fieldNorm(doc=5671)
      0.25 = coord(1/4)
    
    Abstract
    Recommender systems may accelerate knowledge discovery in many fields. However, their users may be competitors guarding their ideas before publication or for other reasons. We describe a simulation experiment to assess user privacy against targeted attacks, modeling recommendations based on co-access data. The analysis uses an unusually long (14?years) set of anonymized historical data on user-item accesses. We introduce the notions of "visibility" and "discoverability." We find, based on historical data, that the majority of the actions of arXiv users would be potentially "visible" under targeted attack. However, "discoverability," which incorporates the difficulty of actually seeing a "visible" effect, is very much lower for nearly all users. We consider the effect of changes to the settings of the recommender algorithm on the visibility and discoverability of user actions and propose mitigation strategies that reduce both measures of risk.
  16. Parapar, J.; Losada, D.E.; Presedo-Quindimil, M.A.; Barreiro, A.: Using score distributions to compare statistical significance tests for information retrieval evaluation (2020) 0.03
    0.030010901 = product of:
      0.120043606 = sum of:
        0.120043606 = weight(_text_:assess in 5506) [ClassicSimilarity], result of:
          0.120043606 = score(doc=5506,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.32564306 = fieldWeight in 5506, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5506)
      0.25 = coord(1/4)
    
    Abstract
    Statistical significance tests can provide evidence that the observed difference in performance between 2 methods is not due to chance. In information retrieval (IR), some studies have examined the validity and suitability of such tests for comparing search systems. We argue here that current methods for assessing the reliability of statistical tests suffer from some methodological weaknesses, and we propose a novel way to study significance tests for retrieval evaluation. Using Score Distributions, we model the output of multiple search systems, produce simulated search results from such models, and compare them using various significance tests. A key strength of this approach is that we assess statistical tests under perfect knowledge about the truth or falseness of the null hypothesis. This new method for studying the power of significance tests in IR evaluation is formal and innovative. Following this type of analysis, we found that both the sign test and Wilcoxon signed test have more power than the permutation test and the t-test. The sign test and Wilcoxon signed test also have good behavior in terms of type I errors. The bootstrap test shows few type I errors, but it has less power than the other methods tested.
  17. Pekar, V.; Binner, J.; Najafi, H.: Early detection of heterogeneous disaster events using social media (2020) 0.03
    0.030010901 = product of:
      0.120043606 = sum of:
        0.120043606 = weight(_text_:assess in 5507) [ClassicSimilarity], result of:
          0.120043606 = score(doc=5507,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.32564306 = fieldWeight in 5507, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5507)
      0.25 = coord(1/4)
    
    Abstract
    This article addresses the problem of detecting crisis-related messages on social media, in order to improve the situational awareness of emergency services. Previous work focused on developing machine-learning classifiers restricted to specific disasters, such as storms or wildfires. We investigate for the first time methods to detect such messages where the type of the crisis is not known in advance, that is, the data are highly heterogeneous. Data heterogeneity causes significant difficulties for learning algorithms to generalize and accurately label incoming data. Our main contributions are as follows. First, we evaluate the extent of this problem in the context of disaster management, finding that the performance of traditional learners drops by up to 40% when trained and tested on heterogeneous data vis-á-vis homogeneous data. Then, in order to overcome data heterogeneity, we propose a new ensemble learning method, and found this to perform on a par with the Gradient Boosting and AdaBoost ensemble learners. The methods are studied on a benchmark data set comprising 26 disaster events and four classification problems: detection of relevant messages, informative messages, eyewitness reports, and topical classification of messages. Finally, in a case study, we evaluate the proposed methods on a real-world data set to assess its practical value.
  18. Wang, H.; Song, Y.-Q.; Wang, L.-T.: Memory model for web ad effect based on multimodal features (2020) 0.03
    0.030010901 = product of:
      0.120043606 = sum of:
        0.120043606 = weight(_text_:assess in 5512) [ClassicSimilarity], result of:
          0.120043606 = score(doc=5512,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.32564306 = fieldWeight in 5512, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5512)
      0.25 = coord(1/4)
    
    Abstract
    Web ad effect evaluation is a challenging problem in web marketing research. Although the analysis of web ad effectiveness has achieved excellent results, there are still some deficiencies. First, there is a lack of an in-depth study of the relevance between advertisements and web content. Second, there is not a thorough analysis of the impacts of users and advertising features on user browsing behaviors. And last, the evaluation index of the web advertisement effect is not adequate. Given the above problems, we conducted our work by studying the observer's behavioral pattern based on multimodal features. First, we analyze the correlation between ads and links with different searching results and further assess the influence of relevance on the observer's attention to web ads using eye-movement features. Then we investigate the user's behavioral sequence and propose the directional frequent-browsing pattern algorithm for mining the user's most commonly used browsing patterns. Finally, we offer the novel use of "memory" as a new measure of advertising effectiveness and further build an advertising memory model with integrated multimodal features for predicting the efficacy of web ads. A large number of experiments have proved the superiority of our method.
  19. Zimmerman, M.S.: Mapping literacies : comparing information horizons mapping to measures of information and health literacy (2020) 0.03
    0.030010901 = product of:
      0.120043606 = sum of:
        0.120043606 = weight(_text_:assess in 5711) [ClassicSimilarity], result of:
          0.120043606 = score(doc=5711,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.32564306 = fieldWeight in 5711, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5711)
      0.25 = coord(1/4)
    
    Abstract
    Purpose Information literacy and health literacy skills are positively correlated with indicators of quality of life. Assessing these literacies, however, can be daunting - particularly with people that may not respond well to prose-based tools. The purpose of this paper is to use information horizons methodology as a metric that may be reflective of literacies. Design/methodology/approach Following a power analysis to insure statistical significance, a sample of 161 participants was recruited from a university population and given formal, vetted measures of information literacy and health literacy and then was asked to create an information horizons map within a health-related context. The information horizons maps were evaluated in two different ways. First, the number of sources was counted. Then, the quality of sources was factored in. Multiple regression analysis was applied to both metrics as independent variables with the other assessments as dependent variables. Anker, Reinhart, and Feeley's model provided the conceptual framework for the study. Findings Information horizons mapping was not found to have a significant relationship with measures of information literacy. However, there were strong, statistically significant relationships with the measures of health literacy employed in this study. Originality/value Employing information horizons methodology as a means of providing a metric to assess literacies may be helpful in providing a more complete picture of a person's abilities. While the current assessment tools have value, this method has the potential to provide important information about the health literacy of people who are not traditionally well represented by prose-based measures.
  20. Soares-Silva, D.; Salati Marcondes de Moraes, G.H.; Cappellozza, A.; Morini, C.: Explaining library user loyalty through perceived service quality : what is wrong? (2020) 0.03
    0.030010901 = product of:
      0.120043606 = sum of:
        0.120043606 = weight(_text_:assess in 5951) [ClassicSimilarity], result of:
          0.120043606 = score(doc=5951,freq=2.0), product of:
            0.36863554 = queryWeight, product of:
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.062536046 = queryNorm
            0.32564306 = fieldWeight in 5951, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.8947687 = idf(docFreq=330, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5951)
      0.25 = coord(1/4)
    
    Abstract
    This study validates the adaptation of a loyalty scale for the library scenario and recovers the hierarchical nature of the perceived service quality (PSQ) by operationalizing it as a second-order level construct, composed by the determinants of service quality (DSQ) identified by Parasuraman, Zeithaml, and Berry in 1985. Our hypothesis was that DSQ are distinct and complementary dimensions, in opposition to the overlapping of DSQ proposed in the SERVQUAL and LibQUAL+® models. In addition, the influence of PSQ on user loyalty (UL) was investigated. Using structural equation modeling, we analyzed the survey data of 1,028 users of a network of academic libraries and report 2 main findings. First, it was shown that the 10 DSQ are statistically significant for the evaluation of PSQ. Second, we demonstrated the positive effect of PSQ for UL. The model presented may be used as a diagnostic and benchmarking tool for managers, coordinators, and librarians who seek to evaluate and/or assess the quality of the services offered by their libraries, as well as to identify and/or manage the loyalty level of their users.

Languages

  • e 96
  • d 29

Types

  • a 119
  • el 21
  • m 2
  • p 2
  • x 1
  • More… Less…