Search (1208 results, page 1 of 61)

  • × year_i:[2020 TO 2030}
  1. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.10
    0.09748046 = product of:
      0.19496092 = sum of:
        0.047273763 = product of:
          0.14182128 = sum of:
            0.14182128 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
              0.14182128 = score(doc=862,freq=2.0), product of:
                0.25234294 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.029764405 = queryNorm
                0.56201804 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=862)
          0.33333334 = coord(1/3)
        0.0058658705 = weight(_text_:a in 862) [ClassicSimilarity], result of:
          0.0058658705 = score(doc=862,freq=10.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.1709182 = fieldWeight in 862, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
        0.14182128 = weight(_text_:2f in 862) [ClassicSimilarity], result of:
          0.14182128 = score(doc=862,freq=2.0), product of:
            0.25234294 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.029764405 = queryNorm
            0.56201804 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
      0.5 = coord(3/6)
    
    Abstract
    This research revisits the classic Turing test and compares recent large language models such as ChatGPT for their abilities to reproduce human-level comprehension and compelling text generation. Two task challenges- summary and question answering- prompt ChatGPT to produce original content (98-99%) from a single text entry and sequential questions initially posed by Turing in 1950. We score the original and generated content against the OpenAI GPT-2 Output Detector from 2019, and establish multiple cases where the generated content proves original and undetectable (98%). The question of a machine fooling a human judge recedes in this work relative to the question of "how would one prove it?" The original contribution of the work presents a metric and simple grammatical set for understanding the writing mechanics of chatbots in evaluating their readability and statistical clarity, engagement, delivery, overall quality, and plagiarism risks. While Turing's original prose scores at least 14% below the machine-generated output, whether an algorithm displays hints of Turing's true initial thoughts (the "Lovelace 2.0" test) remains unanswerable.
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
    Type
    a
  2. Kuessner, T.: Logbuch Mathematik (2020) 0.09
    0.087463796 = product of:
      0.2623914 = sum of:
        0.0026232968 = weight(_text_:a in 5956) [ClassicSimilarity], result of:
          0.0026232968 = score(doc=5956,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.07643694 = fieldWeight in 5956, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=5956)
        0.2597681 = weight(_text_:kurt in 5956) [ClassicSimilarity], result of:
          0.2597681 = score(doc=5956,freq=8.0), product of:
            0.24148957 = queryWeight, product of:
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.029764405 = queryNorm
            1.0756907 = fieldWeight in 5956, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.046875 = fieldNorm(doc=5956)
      0.33333334 = coord(2/6)
    
    Abstract
    Ein Beitrag zum Einbürgerschaftsgespräch mit Kurt Gödel und der Verarbeitung in Daniel Kehlmann: Geister in Princeton.
    Biographed
    Gödel, Kurt
    Content
    Kurt Gödels Staatsbürgerschaftstest "Nun, Mister Gödel, wo kommen Sie her?" "Wo ich herkomme? Österreich." "Was für eine Regierung hatten sie in Österreich?" "Es war eine Republik, doch die Verfassung war so, dass sie in eine Diktatur verwandelt wurde." "Oh, das ist schlecht. Das kann in diesem Land nicht passieren." "Aber ja. Ich kann es beweisen." "Oh Gott. Lassen wir uns da nicht ins Detail gehen." Dieses Gespräch führte Kurt Gödel 1947 bei seinem Einbürgerungstest. Zur Vorbereitung auf die Prüfung hatte er in den Monaten zuvor die Verfassung der USA gelesen. Dem Spieltheoretiker Oskar Morgenstern hatte er danach aufgeregt von inneren Widersprüchen in der Verfassung erzählt und dass es auf völlig legale Weise möglich sei, "ein Diktator zu werden und ein faschistisches Regime zu installieren". Um Probleme beim Einbürgerungsgespräch zu vermeiden, begleiteten ihn deshalb Morgenstern und Albert Einstein, die als Bürgen - anders als sonst üblich - mit ihm gemeinsam befragt wurden..
    Type
    a
  3. Gabler, S.: Vergabe von DDC-Sachgruppen mittels eines Schlagwort-Thesaurus (2021) 0.05
    0.052526407 = product of:
      0.15757921 = sum of:
        0.039394803 = product of:
          0.11818441 = sum of:
            0.11818441 = weight(_text_:3a in 1000) [ClassicSimilarity], result of:
              0.11818441 = score(doc=1000,freq=2.0), product of:
                0.25234294 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.029764405 = queryNorm
                0.46834838 = fieldWeight in 1000, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1000)
          0.33333334 = coord(1/3)
        0.11818441 = weight(_text_:2f in 1000) [ClassicSimilarity], result of:
          0.11818441 = score(doc=1000,freq=2.0), product of:
            0.25234294 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.029764405 = queryNorm
            0.46834838 = fieldWeight in 1000, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1000)
      0.33333334 = coord(2/6)
    
    Content
    Master thesis Master of Science (Library and Information Studies) (MSc), Universität Wien. Advisor: Christoph Steiner. Vgl.: https://www.researchgate.net/publication/371680244_Vergabe_von_DDC-Sachgruppen_mittels_eines_Schlagwort-Thesaurus. DOI: 10.25365/thesis.70030. Vgl. dazu die Präsentation unter: https://www.google.com/url?sa=i&rct=j&q=&esrc=s&source=web&cd=&ved=0CAIQw7AJahcKEwjwoZzzytz_AhUAAAAAHQAAAAAQAg&url=https%3A%2F%2Fwiki.dnb.de%2Fdownload%2Fattachments%2F252121510%2FDA3%2520Workshop-Gabler.pdf%3Fversion%3D1%26modificationDate%3D1671093170000%26api%3Dv2&psig=AOvVaw0szwENK1or3HevgvIDOfjx&ust=1687719410889597&opi=89978449.
  4. Zilm, G.: "Kl ist ein glorifizierter Taschenrechner" (2023) 0.01
    0.014706166 = product of:
      0.029412333 = sum of:
        0.0043721613 = weight(_text_:a in 1129) [ClassicSimilarity], result of:
          0.0043721613 = score(doc=1129,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.12739488 = fieldWeight in 1129, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.078125 = fieldNorm(doc=1129)
        0.01159795 = product of:
          0.0463918 = sum of:
            0.0463918 = weight(_text_:g in 1129) [ClassicSimilarity], result of:
              0.0463918 = score(doc=1129,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.4149775 = fieldWeight in 1129, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1129)
          0.25 = coord(1/4)
        0.013442221 = product of:
          0.040326662 = sum of:
            0.040326662 = weight(_text_:22 in 1129) [ClassicSimilarity], result of:
              0.040326662 = score(doc=1129,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.38690117 = fieldWeight in 1129, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1129)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Date
    27. 1.2023 16:22:55
    Type
    a
  5. Candela, G.: ¬An automatic data quality approach to assess semantic data from cultural heritage institutions (2023) 0.01
    0.010928169 = product of:
      0.021856338 = sum of:
        0.004328219 = weight(_text_:a in 997) [ClassicSimilarity], result of:
          0.004328219 = score(doc=997,freq=4.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.12611452 = fieldWeight in 997, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=997)
        0.008118565 = product of:
          0.03247426 = sum of:
            0.03247426 = weight(_text_:g in 997) [ClassicSimilarity], result of:
              0.03247426 = score(doc=997,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.29048425 = fieldWeight in 997, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=997)
          0.25 = coord(1/4)
        0.009409554 = product of:
          0.028228661 = sum of:
            0.028228661 = weight(_text_:22 in 997) [ClassicSimilarity], result of:
              0.028228661 = score(doc=997,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.2708308 = fieldWeight in 997, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=997)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Abstract
    In recent years, cultural heritage institutions have been exploring the benefits of applying Linked Open Data to their catalogs and digital materials. Innovative and creative methods have emerged to publish and reuse digital contents to promote computational access, such as the concepts of Labs and Collections as Data. Data quality has become a requirement for researchers and training methods based on artificial intelligence and machine learning. This article explores how the quality of Linked Open Data made available by cultural heritage institutions can be automatically assessed. The results obtained can be useful for other institutions who wish to publish and assess their collections.
    Date
    22. 6.2023 18:23:31
    Type
    a
  6. Geras, A.; Siudem, G.; Gagolewski, M.: Should we introduce a dislike button for academic articles? (2020) 0.01
    0.010444986 = product of:
      0.020889971 = sum of:
        0.0058658705 = weight(_text_:a in 5620) [ClassicSimilarity], result of:
          0.0058658705 = score(doc=5620,freq=10.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.1709182 = fieldWeight in 5620, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=5620)
        0.0069587696 = product of:
          0.027835079 = sum of:
            0.027835079 = weight(_text_:g in 5620) [ClassicSimilarity], result of:
              0.027835079 = score(doc=5620,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.24898648 = fieldWeight in 5620, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5620)
          0.25 = coord(1/4)
        0.008065332 = product of:
          0.024195995 = sum of:
            0.024195995 = weight(_text_:22 in 5620) [ClassicSimilarity], result of:
              0.024195995 = score(doc=5620,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.23214069 = fieldWeight in 5620, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5620)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Abstract
    There is a mutual resemblance between the behavior of users of the Stack Exchange and the dynamics of the citations accumulation process in the scientific community, which enabled us to tackle the outwardly intractable problem of assessing the impact of introducing "negative" citations. Although the most frequent reason to cite an article is to highlight the connection between the 2 publications, researchers sometimes mention an earlier work to cast a negative light. While computing citation-based scores, for instance, the h-index, information about the reason why an article was mentioned is neglected. Therefore, it can be questioned whether these indices describe scientific achievements accurately. In this article we shed insight into the problem of "negative" citations, analyzing data from Stack Exchange and, to draw more universal conclusions, we derive an approximation of citations scores. Here we show that the quantified influence of introducing negative citations is of lesser importance and that they could be used as an indicator of where the attention of the scientific community is allocated.
    Date
    6. 1.2020 18:10:22
    Type
    a
  7. Vogt, T.: ¬Die Transformation des renommierten Informationsservices zbMATH zu einer Open Access-Plattform für die Mathematik steht vor dem Abschluss. (2020) 0.01
    0.010422628 = product of:
      0.06253577 = sum of:
        0.06253577 = product of:
          0.25014308 = sum of:
            0.25014308 = weight(_text_:c3 in 31) [ClassicSimilarity], result of:
              0.25014308 = score(doc=31,freq=2.0), product of:
                0.29023203 = queryWeight, product of:
                  9.7509775 = idf(docFreq=6, maxDocs=44218)
                  0.029764405 = queryNorm
                0.8618728 = fieldWeight in 31, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  9.7509775 = idf(docFreq=6, maxDocs=44218)
                  0.0625 = fieldNorm(doc=31)
          0.25 = coord(1/4)
      0.16666667 = coord(1/6)
    
    Content
    "Mit Beginn des Jahres 2021 wird der umfassende internationale Informationsservice zbMATH in eine Open Access-Plattform überführt. Dann steht dieser bislang kostenpflichtige Dienst weltweit allen Interessierten kostenfrei zur Verfügung. Die Änderung des Geschäftsmodells ermöglicht, die meisten Informationen und Daten von zbMATH für Forschungszwecke und zur Verknüpfung mit anderen nicht-kommerziellen Diensten frei zu nutzen, siehe: https://www.mathematik.de/dmv-blog/2772-transformation-von-zbmath-zu-einer-open-access-plattform-f%C3%BCr-die-mathematik-kurz-vor-dem-abschluss."
  8. Li, G.; Siddharth, L.; Luo, J.: Embedding knowledge graph of patent metadata to measure knowledge proximity (2023) 0.01
    0.010135347 = product of:
      0.020270694 = sum of:
        0.0052465936 = weight(_text_:a in 920) [ClassicSimilarity], result of:
          0.0052465936 = score(doc=920,freq=8.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.15287387 = fieldWeight in 920, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=920)
        0.0069587696 = product of:
          0.027835079 = sum of:
            0.027835079 = weight(_text_:g in 920) [ClassicSimilarity], result of:
              0.027835079 = score(doc=920,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.24898648 = fieldWeight in 920, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.046875 = fieldNorm(doc=920)
          0.25 = coord(1/4)
        0.008065332 = product of:
          0.024195995 = sum of:
            0.024195995 = weight(_text_:22 in 920) [ClassicSimilarity], result of:
              0.024195995 = score(doc=920,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.23214069 = fieldWeight in 920, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=920)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Abstract
    Knowledge proximity refers to the strength of association between any two entities in a structural form that embodies certain aspects of a knowledge base. In this work, we operationalize knowledge proximity within the context of the US Patent Database (knowledge base) using a knowledge graph (structural form) named "PatNet" built using patent metadata, including citations, inventors, assignees, and domain classifications. We train various graph embedding models using PatNet to obtain the embeddings of entities and relations. The cosine similarity between the corresponding (or transformed) embeddings of entities denotes the knowledge proximity between these. We compare the embedding models in terms of their performances in predicting target entities and explaining domain expansion profiles of inventors and assignees. We then apply the embeddings of the best-preferred model to associate homogeneous (e.g., patent-patent) and heterogeneous (e.g., inventor-assignee) pairs of entities.
    Date
    22. 3.2023 12:06:55
    Type
    a
  9. Zhang, Y.; Wu, M.; Zhang, G.; Lu, J.: Stepping beyond your comfort zone : diffusion-based network analytics for knowledge trajectory recommendation (2023) 0.01
    0.009885248 = product of:
      0.019770496 = sum of:
        0.00725041 = weight(_text_:a in 994) [ClassicSimilarity], result of:
          0.00725041 = score(doc=994,freq=22.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.21126054 = fieldWeight in 994, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=994)
        0.005798975 = product of:
          0.0231959 = sum of:
            0.0231959 = weight(_text_:g in 994) [ClassicSimilarity], result of:
              0.0231959 = score(doc=994,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.20748875 = fieldWeight in 994, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=994)
          0.25 = coord(1/4)
        0.0067211105 = product of:
          0.020163331 = sum of:
            0.020163331 = weight(_text_:22 in 994) [ClassicSimilarity], result of:
              0.020163331 = score(doc=994,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.19345059 = fieldWeight in 994, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=994)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Abstract
    Predicting a researcher's knowledge trajectories beyond their current foci can leverage potential inter-/cross-/multi-disciplinary interactions to achieve exploratory innovation. In this study, we present a method of diffusion-based network analytics for knowledge trajectory recommendation. The method begins by constructing a heterogeneous bibliometric network consisting of a co-topic layer and a co-authorship layer. A novel link prediction approach with a diffusion strategy is then used to capture the interactions between social elements (e.g., collaboration) and knowledge elements (e.g., technological similarity) in the process of exploratory innovation. This diffusion strategy differentiates the interactions occurring among homogeneous and heterogeneous nodes in the heterogeneous bibliometric network and weights the strengths of these interactions. Two sets of experiments-one with a local dataset and the other with a global dataset-demonstrate that the proposed method is prior to 10 selected baselines in link prediction, recommender systems, and upstream graph representation learning. A case study recommending knowledge trajectories of information scientists with topical hierarchy and explainable mediators reveals the proposed method's reliability and potential practical uses in broad scenarios.
    Date
    22. 6.2023 18:07:12
    Type
    a
  10. Cerda-Cosme, R.; Méndez, E.: Analysis of shared research data in Spanish scientific papers about COVID-19 : a first approach (2023) 0.01
    0.008937434 = product of:
      0.017874869 = sum of:
        0.0053547826 = weight(_text_:a in 916) [ClassicSimilarity], result of:
          0.0053547826 = score(doc=916,freq=12.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.15602624 = fieldWeight in 916, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=916)
        0.005798975 = product of:
          0.0231959 = sum of:
            0.0231959 = weight(_text_:g in 916) [ClassicSimilarity], result of:
              0.0231959 = score(doc=916,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.20748875 = fieldWeight in 916, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=916)
          0.25 = coord(1/4)
        0.0067211105 = product of:
          0.020163331 = sum of:
            0.020163331 = weight(_text_:22 in 916) [ClassicSimilarity], result of:
              0.020163331 = score(doc=916,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.19345059 = fieldWeight in 916, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=916)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Abstract
    During the coronavirus pandemic, changes in the way science is done and shared occurred, which motivates meta-research to help understand science communication in crises and improve its effectiveness. The objective is to study how many Spanish scientific papers on COVID-19 published during 2020 share their research data. Qualitative and descriptive study applying nine attributes: (a) availability, (b) accessibility, (c) format, (d) licensing, (e) linkage, (f) funding, (g) editorial policy, (h) content, and (i) statistics. We analyzed 1,340 papers, 1,173 (87.5%) did not have research data. A total of 12.5% share their research data of which 2.1% share their data in repositories, 5% share their data through a simple request, 0.2% do not have permission to share their data, and 5.2% share their data as supplementary material. There is a small percentage that shares their research data; however, it demonstrates the researchers' poor knowledge on how to properly share their research data and their lack of knowledge on what is research data.
    Date
    21. 3.2023 19:22:02
    Type
    a
  11. Hindrichs, G.: Kriegszivilgesellschaft (2022) 0.01
    0.008517393 = product of:
      0.025552178 = sum of:
        0.006995458 = weight(_text_:a in 986) [ClassicSimilarity], result of:
          0.006995458 = score(doc=986,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.20383182 = fieldWeight in 986, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.125 = fieldNorm(doc=986)
        0.01855672 = product of:
          0.07422688 = sum of:
            0.07422688 = weight(_text_:g in 986) [ClassicSimilarity], result of:
              0.07422688 = score(doc=986,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.663964 = fieldWeight in 986, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.125 = fieldNorm(doc=986)
          0.25 = coord(1/4)
      0.33333334 = coord(2/6)
    
    Type
    a
  12. Wang, S.; Ma, Y.; Mao, J.; Bai, Y.; Liang, Z.; Li, G.: Quantifying scientific breakthroughs by a novel disruption indicator based on knowledge entities : On the rise of scrape-and-report scholarship in online reviews research (2023) 0.01
    0.008446123 = product of:
      0.016892247 = sum of:
        0.0043721613 = weight(_text_:a in 882) [ClassicSimilarity], result of:
          0.0043721613 = score(doc=882,freq=8.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.12739488 = fieldWeight in 882, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=882)
        0.005798975 = product of:
          0.0231959 = sum of:
            0.0231959 = weight(_text_:g in 882) [ClassicSimilarity], result of:
              0.0231959 = score(doc=882,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.20748875 = fieldWeight in 882, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=882)
          0.25 = coord(1/4)
        0.0067211105 = product of:
          0.020163331 = sum of:
            0.020163331 = weight(_text_:22 in 882) [ClassicSimilarity], result of:
              0.020163331 = score(doc=882,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.19345059 = fieldWeight in 882, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=882)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Abstract
    Compared to previous studies that generally detect scientific breakthroughs based on citation patterns, this article proposes a knowledge entity-based disruption indicator by quantifying the change of knowledge directly created and inspired by scientific breakthroughs to their evolutionary trajectories. Two groups of analytic units, including MeSH terms and their co-occurrences, are employed independently by the indicator to measure the change of knowledge. The effectiveness of the proposed indicators was evaluated against the four datasets of scientific breakthroughs derived from four recognition trials. In terms of identifying scientific breakthroughs, the proposed disruption indicator based on MeSH co-occurrences outperforms that based on MeSH terms and three earlier citation-based disruption indicators. It is also shown that in our indicator, measuring the change of knowledge inspired by the focal paper in its evolutionary trajectory is a larger contributor than measuring the change created by the focal paper. Our study not only offers empirical insights into conceptual understanding of scientific breakthroughs but also provides practical disruption indicator for scientists and science management agencies searching for valuable research.
    Date
    22. 1.2023 18:37:33
    Type
    a
  13. Yu, C.; Xue, H.; An, L.; Li, G.: ¬A lightweight semantic-enhanced interactive network for efficient short-text matching (2023) 0.01
    0.008446123 = product of:
      0.016892247 = sum of:
        0.0043721613 = weight(_text_:a in 890) [ClassicSimilarity], result of:
          0.0043721613 = score(doc=890,freq=8.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.12739488 = fieldWeight in 890, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=890)
        0.005798975 = product of:
          0.0231959 = sum of:
            0.0231959 = weight(_text_:g in 890) [ClassicSimilarity], result of:
              0.0231959 = score(doc=890,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.20748875 = fieldWeight in 890, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=890)
          0.25 = coord(1/4)
        0.0067211105 = product of:
          0.020163331 = sum of:
            0.020163331 = weight(_text_:22 in 890) [ClassicSimilarity], result of:
              0.020163331 = score(doc=890,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.19345059 = fieldWeight in 890, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=890)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Abstract
    Knowledge-enhanced short-text matching has been a significant task attracting much attention in recent years. However, the existing approaches cannot effectively balance effect and efficiency. Effective models usually consist of complex network structures leading to slow inference speed and the difficulties of applications in actual practice. In addition, most knowledge-enhanced models try to link the mentions in the text to the entities of the knowledge graphs-the difficulties of entity linking decrease the generalizability among different datasets. To address these problems, we propose a lightweight Semantic-Enhanced Interactive Network (SEIN) model for efficient short-text matching. Unlike most current research, SEIN employs an unsupervised method to select WordNet's most appropriate paraphrase description as the external semantic knowledge. It focuses on integrating semantic information and interactive information of text while simplifying the structure of other modules. We conduct intensive experiments on four real-world datasets, that is, Quora, Twitter-URL, SciTail, and SICK-E. Compared with state-of-the-art methods, SEIN achieves the best performance on most datasets. The experimental results proved that introducing external knowledge could effectively improve the performance of the short-text matching models. The research sheds light on the role of lightweight models in leveraging external knowledge to improve the effect of short-text matching.
    Date
    22. 1.2023 19:05:27
    Type
    a
  14. Luo, L.; Ju, J.; Li, Y.-F.; Haffari, G.; Xiong, B.; Pan, S.: ChatRule: mining logical rules with large language models for knowledge graph reasoning (2023) 0.01
    0.008446123 = product of:
      0.016892247 = sum of:
        0.0043721613 = weight(_text_:a in 1171) [ClassicSimilarity], result of:
          0.0043721613 = score(doc=1171,freq=8.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.12739488 = fieldWeight in 1171, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1171)
        0.005798975 = product of:
          0.0231959 = sum of:
            0.0231959 = weight(_text_:g in 1171) [ClassicSimilarity], result of:
              0.0231959 = score(doc=1171,freq=2.0), product of:
                0.11179353 = queryWeight, product of:
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.029764405 = queryNorm
                0.20748875 = fieldWeight in 1171, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.7559474 = idf(docFreq=2809, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1171)
          0.25 = coord(1/4)
        0.0067211105 = product of:
          0.020163331 = sum of:
            0.020163331 = weight(_text_:22 in 1171) [ClassicSimilarity], result of:
              0.020163331 = score(doc=1171,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.19345059 = fieldWeight in 1171, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1171)
          0.33333334 = coord(1/3)
      0.5 = coord(3/6)
    
    Abstract
    Logical rules are essential for uncovering the logical connections between relations, which could improve the reasoning performance and provide interpretable results on knowledge graphs (KGs). Although there have been many efforts to mine meaningful logical rules over KGs, existing methods suffer from the computationally intensive searches over the rule space and a lack of scalability for large-scale KGs. Besides, they often ignore the semantics of relations which is crucial for uncovering logical connections. Recently, large language models (LLMs) have shown impressive performance in the field of natural language processing and various applications, owing to their emergent ability and generalizability. In this paper, we propose a novel framework, ChatRule, unleashing the power of large language models for mining logical rules over knowledge graphs. Specifically, the framework is initiated with an LLM-based rule generator, leveraging both the semantic and structural information of KGs to prompt LLMs to generate logical rules. To refine the generated rules, a rule ranking module estimates the rule quality by incorporating facts from existing KGs. Last, a rule validator harnesses the reasoning ability of LLMs to validate the logical correctness of ranked rules through chain-of-thought reasoning. ChatRule is evaluated on four large-scale KGs, w.r.t. different rule quality metrics and downstream tasks, showing the effectiveness and scalability of our method.
    Date
    23.11.2023 19:07:22
  15. ¬Der Student aus dem Computer (2023) 0.01
    0.008313378 = product of:
      0.024940135 = sum of:
        0.006121026 = weight(_text_:a in 1079) [ClassicSimilarity], result of:
          0.006121026 = score(doc=1079,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.17835285 = fieldWeight in 1079, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=1079)
        0.018819109 = product of:
          0.056457322 = sum of:
            0.056457322 = weight(_text_:22 in 1079) [ClassicSimilarity], result of:
              0.056457322 = score(doc=1079,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.5416616 = fieldWeight in 1079, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1079)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Date
    27. 1.2023 16:22:55
    Type
    a
  16. Jaeger, L.: Wissenschaftler versus Wissenschaft (2020) 0.01
    0.007125753 = product of:
      0.021377258 = sum of:
        0.0052465936 = weight(_text_:a in 4156) [ClassicSimilarity], result of:
          0.0052465936 = score(doc=4156,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.15287387 = fieldWeight in 4156, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.09375 = fieldNorm(doc=4156)
        0.016130663 = product of:
          0.04839199 = sum of:
            0.04839199 = weight(_text_:22 in 4156) [ClassicSimilarity], result of:
              0.04839199 = score(doc=4156,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.46428138 = fieldWeight in 4156, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4156)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Date
    2. 3.2020 14:08:22
    Type
    a
  17. Ibrahim, G.M.; Taylor, M.: Krebszellen manipulieren Neurone : Gliome (2023) 0.01
    0.007125753 = product of:
      0.021377258 = sum of:
        0.0052465936 = weight(_text_:a in 1203) [ClassicSimilarity], result of:
          0.0052465936 = score(doc=1203,freq=2.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.15287387 = fieldWeight in 1203, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.09375 = fieldNorm(doc=1203)
        0.016130663 = product of:
          0.04839199 = sum of:
            0.04839199 = weight(_text_:22 in 1203) [ClassicSimilarity], result of:
              0.04839199 = score(doc=1203,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.46428138 = fieldWeight in 1203, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=1203)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Source
    Spektrum der Wissenschaft. 2023, H.10, S.22-24
    Type
    a
  18. Dietz, K.: en.wikipedia.org > 6 Mio. Artikel (2020) 0.01
    0.006565801 = product of:
      0.039394803 = sum of:
        0.039394803 = product of:
          0.11818441 = sum of:
            0.11818441 = weight(_text_:3a in 5669) [ClassicSimilarity], result of:
              0.11818441 = score(doc=5669,freq=2.0), product of:
                0.25234294 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.029764405 = queryNorm
                0.46834838 = fieldWeight in 5669, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5669)
          0.33333334 = coord(1/3)
      0.16666667 = coord(1/6)
    
    Content
    "Die Englischsprachige Wikipedia verfügt jetzt über mehr als 6 Millionen Artikel. An zweiter Stelle kommt die deutschsprachige Wikipedia mit 2.3 Millionen Artikeln, an dritter Stelle steht die französischsprachige Wikipedia mit 2.1 Millionen Artikeln (via Researchbuzz: Firehose <https://rbfirehose.com/2020/01/24/techcrunch-wikipedia-now-has-more-than-6-million-articles-in-english/> und Techcrunch <https://techcrunch.com/2020/01/23/wikipedia-english-six-million-articles/?utm_source=feedburner&utm_medium=feed&utm_campaign=Feed%3A+Techcrunch+%28TechCrunch%29&guccounter=1&guce_referrer=aHR0cHM6Ly9yYmZpcmVob3NlLmNvbS8yMDIwLzAxLzI0L3RlY2hjcnVuY2gtd2lraXBlZGlhLW5vdy1oYXMtbW9yZS10aGFuLTYtbWlsbGlvbi1hcnRpY2xlcy1pbi1lbmdsaXNoLw&guce_referrer_sig=AQAAAK0zHfjdDZ_spFZBF_z-zDjtL5iWvuKDumFTzm4HvQzkUfE2pLXQzGS6FGB_y-VISdMEsUSvkNsg2U_NWQ4lwWSvOo3jvXo1I3GtgHpP8exukVxYAnn5mJspqX50VHIWFADHhs5AerkRn3hMRtf_R3F1qmEbo8EROZXp328HMC-o>). 250120 via digithek ch = #fineBlog s.a.: Angesichts der Veröffentlichung des 6-millionsten Artikels vergangene Woche in der englischsprachigen Wikipedia hat die Community-Zeitungsseite "Wikipedia Signpost" ein Moratorium bei der Veröffentlichung von Unternehmensartikeln gefordert. Das sei kein Vorwurf gegen die Wikimedia Foundation, aber die derzeitigen Maßnahmen, um die Enzyklopädie gegen missbräuchliches undeklariertes Paid Editing zu schützen, funktionierten ganz klar nicht. *"Da die ehrenamtlichen Autoren derzeit von Werbung in Gestalt von Wikipedia-Artikeln überwältigt werden, und da die WMF nicht in der Lage zu sein scheint, dem irgendetwas entgegenzusetzen, wäre der einzige gangbare Weg für die Autoren, fürs erste die Neuanlage von Artikeln über Unternehmen zu untersagen"*, schreibt der Benutzer Smallbones in seinem Editorial <https://en.wikipedia.org/wiki/Wikipedia:Wikipedia_Signpost/2020-01-27/From_the_editor> zur heutigen Ausgabe."
  19. Sokolow, A.: Es menschelt in der KI-Welt (2023) 0.01
    0.0065417974 = product of:
      0.019625392 = sum of:
        0.0061831702 = weight(_text_:a in 1169) [ClassicSimilarity], result of:
          0.0061831702 = score(doc=1169,freq=4.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.18016359 = fieldWeight in 1169, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.078125 = fieldNorm(doc=1169)
        0.013442221 = product of:
          0.040326662 = sum of:
            0.040326662 = weight(_text_:22 in 1169) [ClassicSimilarity], result of:
              0.040326662 = score(doc=1169,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.38690117 = fieldWeight in 1169, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1169)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Date
    27. 1.2023 16:22:55
    Type
    a
  20. Sokolow, A.: Chaostage bei ChatGPT (2023) 0.01
    0.0065417974 = product of:
      0.019625392 = sum of:
        0.0061831702 = weight(_text_:a in 1170) [ClassicSimilarity], result of:
          0.0061831702 = score(doc=1170,freq=4.0), product of:
            0.034319755 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.029764405 = queryNorm
            0.18016359 = fieldWeight in 1170, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.078125 = fieldNorm(doc=1170)
        0.013442221 = product of:
          0.040326662 = sum of:
            0.040326662 = weight(_text_:22 in 1170) [ClassicSimilarity], result of:
              0.040326662 = score(doc=1170,freq=2.0), product of:
                0.104229875 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.029764405 = queryNorm
                0.38690117 = fieldWeight in 1170, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1170)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Date
    27. 1.2023 16:22:55
    Type
    a

Languages

  • e 820
  • d 379
  • pt 6
  • sp 1
  • More… Less…

Types

  • a 1148
  • el 247
  • m 29
  • p 13
  • s 7
  • x 2
  • A 1
  • EL 1
  • More… Less…

Themes

Subjects

Classifications