Search (5 results, page 1 of 1)

  • × year_i:[2020 TO 2030}
  • × theme_ss:"Data Mining"
  1. Datentracking in der Wissenschaft : Aggregation und Verwendung bzw. Verkauf von Nutzungsdaten durch Wissenschaftsverlage. Ein Informationspapier des Ausschusses für Wissenschaftliche Bibliotheken und Informationssysteme der Deutschen Forschungsgemeinschaft (2021) 0.04
    0.04414154 = product of:
      0.16553076 = sum of:
        0.024400797 = weight(_text_:und in 248) [ClassicSimilarity], result of:
          0.024400797 = score(doc=248,freq=24.0), product of:
            0.04794192 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021630857 = queryNorm
            0.5089658 = fieldWeight in 248, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=248)
        0.024400797 = weight(_text_:und in 248) [ClassicSimilarity], result of:
          0.024400797 = score(doc=248,freq=24.0), product of:
            0.04794192 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021630857 = queryNorm
            0.5089658 = fieldWeight in 248, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=248)
        0.018930178 = weight(_text_:der in 248) [ClassicSimilarity], result of:
          0.018930178 = score(doc=248,freq=14.0), product of:
            0.04831829 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.021630857 = queryNorm
            0.3917808 = fieldWeight in 248, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.046875 = fieldNorm(doc=248)
        0.024400797 = weight(_text_:und in 248) [ClassicSimilarity], result of:
          0.024400797 = score(doc=248,freq=24.0), product of:
            0.04794192 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021630857 = queryNorm
            0.5089658 = fieldWeight in 248, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=248)
        0.024400797 = weight(_text_:und in 248) [ClassicSimilarity], result of:
          0.024400797 = score(doc=248,freq=24.0), product of:
            0.04794192 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021630857 = queryNorm
            0.5089658 = fieldWeight in 248, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=248)
        0.010834851 = weight(_text_:20 in 248) [ClassicSimilarity], result of:
          0.010834851 = score(doc=248,freq=2.0), product of:
            0.059459355 = queryWeight, product of:
              2.748821 = idf(docFreq=7692, maxDocs=44218)
              0.021630857 = queryNorm
            0.18222281 = fieldWeight in 248, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.748821 = idf(docFreq=7692, maxDocs=44218)
              0.046875 = fieldNorm(doc=248)
        0.024400797 = weight(_text_:und in 248) [ClassicSimilarity], result of:
          0.024400797 = score(doc=248,freq=24.0), product of:
            0.04794192 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021630857 = queryNorm
            0.5089658 = fieldWeight in 248, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.046875 = fieldNorm(doc=248)
        0.013761732 = weight(_text_:21 in 248) [ClassicSimilarity], result of:
          0.013761732 = score(doc=248,freq=2.0), product of:
            0.06701087 = queryWeight, product of:
              3.0979297 = idf(docFreq=5425, maxDocs=44218)
              0.021630857 = queryNorm
            0.20536566 = fieldWeight in 248, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0979297 = idf(docFreq=5425, maxDocs=44218)
              0.046875 = fieldNorm(doc=248)
      0.26666668 = coord(8/30)
    
    Abstract
    Das Informationspapier beschreibt die digitale Nachverfolgung von wissenschaftlichen Aktivitäten. Wissenschaftlerinnen und Wissenschaftler nutzen täglich eine Vielzahl von digitalen Informationsressourcen wie zum Beispiel Literatur- und Volltextdatenbanken. Häufig fallen dabei Nutzungsspuren an, die Aufschluss geben über gesuchte und genutzte Inhalte, Verweildauern und andere Arten der wissenschaftlichen Aktivität. Diese Nutzungsspuren können von den Anbietenden der Informationsressourcen festgehalten, aggregiert und weiterverwendet oder verkauft werden. Das Informationspapier legt die Transformation von Wissenschaftsverlagen hin zu Data Analytics Businesses dar, verweist auf die Konsequenzen daraus für die Wissenschaft und deren Einrichtungen und benennt die zum Einsatz kommenden Typen der Datengewinnung. Damit dient es vor allem der Darstellung gegenwärtiger Praktiken und soll zu Diskussionen über deren Konsequenzen für die Wissenschaft anregen. Es richtet sich an alle Wissenschaftlerinnen und Wissenschaftler sowie alle Akteure in der Wissenschaftslandschaft.
    Date
    21. 5.2021 12:10:03
    Editor
    Deutsche Forschungsgemeinschaft / Ausschuss für Wissenschaftliche Bibliotheken und Informationssysteme
    Issue
    [20. Mai 2021].
  2. Mandl, T.: Text Mining und Data Mining (2023) 0.03
    0.028381787 = product of:
      0.121636234 = sum of:
        0.02012963 = weight(_text_:und in 774) [ClassicSimilarity], result of:
          0.02012963 = score(doc=774,freq=12.0), product of:
            0.04794192 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021630857 = queryNorm
            0.41987535 = fieldWeight in 774, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=774)
        0.02012963 = weight(_text_:und in 774) [ClassicSimilarity], result of:
          0.02012963 = score(doc=774,freq=12.0), product of:
            0.04794192 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021630857 = queryNorm
            0.41987535 = fieldWeight in 774, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=774)
        0.008347424 = weight(_text_:der in 774) [ClassicSimilarity], result of:
          0.008347424 = score(doc=774,freq=2.0), product of:
            0.04831829 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.021630857 = queryNorm
            0.17275909 = fieldWeight in 774, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0546875 = fieldNorm(doc=774)
        0.02012963 = weight(_text_:und in 774) [ClassicSimilarity], result of:
          0.02012963 = score(doc=774,freq=12.0), product of:
            0.04794192 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021630857 = queryNorm
            0.41987535 = fieldWeight in 774, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=774)
        0.02012963 = weight(_text_:und in 774) [ClassicSimilarity], result of:
          0.02012963 = score(doc=774,freq=12.0), product of:
            0.04794192 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021630857 = queryNorm
            0.41987535 = fieldWeight in 774, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=774)
        0.012640659 = weight(_text_:20 in 774) [ClassicSimilarity], result of:
          0.012640659 = score(doc=774,freq=2.0), product of:
            0.059459355 = queryWeight, product of:
              2.748821 = idf(docFreq=7692, maxDocs=44218)
              0.021630857 = queryNorm
            0.21259327 = fieldWeight in 774, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.748821 = idf(docFreq=7692, maxDocs=44218)
              0.0546875 = fieldNorm(doc=774)
        0.02012963 = weight(_text_:und in 774) [ClassicSimilarity], result of:
          0.02012963 = score(doc=774,freq=12.0), product of:
            0.04794192 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021630857 = queryNorm
            0.41987535 = fieldWeight in 774, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=774)
      0.23333333 = coord(7/30)
    
    Abstract
    Text und Data Mining sind ein Bündel von Technologien, die eng mit den Themenfeldern Statistik, Maschinelles Lernen und dem Erkennen von Mustern verbunden sind. Die üblichen Definitionen beziehen eine Vielzahl von verschiedenen Verfahren mit ein, ohne eine exakte Grenze zu ziehen. Data Mining bezeichnet die Suche nach Mustern, Regelmäßigkeiten oder Auffälligkeiten in stark strukturierten und vor allem numerischen Daten. "Any algorithm that enumerates patterns from, or fits models to, data is a data mining algorithm." Numerische Daten und Datenbankinhalte werden als strukturierte Daten bezeichnet. Dagegen gelten Textdokumente in natürlicher Sprache als unstrukturierte Daten.
    Date
    24.11.2022 11:42:20
    Source
    Grundlagen der Informationswissenschaft. Hrsg.: Rainer Kuhlen, Dirk Lewandowski, Wolfgang Semar und Christa Womser-Hacker. 7., völlig neu gefasste Ausg
  3. Wiegmann, S.: Hättest du die Titanic überlebt? : Eine kurze Einführung in das Data Mining mit freier Software (2023) 0.03
    0.025048751 = product of:
      0.10735179 = sum of:
        0.016435774 = weight(_text_:und in 876) [ClassicSimilarity], result of:
          0.016435774 = score(doc=876,freq=8.0), product of:
            0.04794192 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021630857 = queryNorm
            0.34282678 = fieldWeight in 876, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=876)
        0.016435774 = weight(_text_:und in 876) [ClassicSimilarity], result of:
          0.016435774 = score(doc=876,freq=8.0), product of:
            0.04794192 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021630857 = queryNorm
            0.34282678 = fieldWeight in 876, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=876)
        0.01180504 = weight(_text_:der in 876) [ClassicSimilarity], result of:
          0.01180504 = score(doc=876,freq=4.0), product of:
            0.04831829 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.021630857 = queryNorm
            0.24431825 = fieldWeight in 876, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0546875 = fieldNorm(doc=876)
        0.016435774 = weight(_text_:und in 876) [ClassicSimilarity], result of:
          0.016435774 = score(doc=876,freq=8.0), product of:
            0.04794192 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021630857 = queryNorm
            0.34282678 = fieldWeight in 876, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=876)
        0.016435774 = weight(_text_:und in 876) [ClassicSimilarity], result of:
          0.016435774 = score(doc=876,freq=8.0), product of:
            0.04794192 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021630857 = queryNorm
            0.34282678 = fieldWeight in 876, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=876)
        0.013367887 = weight(_text_:im in 876) [ClassicSimilarity], result of:
          0.013367887 = score(doc=876,freq=2.0), product of:
            0.06114581 = queryWeight, product of:
              2.8267863 = idf(docFreq=7115, maxDocs=44218)
              0.021630857 = queryNorm
            0.2186231 = fieldWeight in 876, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.8267863 = idf(docFreq=7115, maxDocs=44218)
              0.0546875 = fieldNorm(doc=876)
        0.016435774 = weight(_text_:und in 876) [ClassicSimilarity], result of:
          0.016435774 = score(doc=876,freq=8.0), product of:
            0.04794192 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.021630857 = queryNorm
            0.34282678 = fieldWeight in 876, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=876)
      0.23333333 = coord(7/30)
    
    Abstract
    Am 10. April 1912 ging Elisabeth Walton Allen an Bord der "Titanic", um ihr Hab und Gut nach England zu holen. Eines Nachts wurde sie von ihrer aufgelösten Tante geweckt, deren Kajüte unter Wasser stand. Wie steht es um Elisabeths Chancen und hätte man selbst das Unglück damals überlebt? Das Titanic-Orakel ist eine algorithmusbasierte App, die entsprechende Prognosen aufstellt und im Rahmen des Kurses "Data Science" am Department Information der HAW Hamburg entstanden ist. Dieser Beitrag zeigt Schritt für Schritt, wie die App unter Verwendung freier Software entwickelt wurde. Code und Daten werden zur Nachnutzung bereitgestellt.
  4. Borgman, C.L.; Wofford, M.F.; Golshan, M.S.; Darch, P.T.: Collaborative qualitative research at scale : reflections on 20 years of acquiring global data and making data global (2021) 0.00
    5.212921E-4 = product of:
      0.015638761 = sum of:
        0.015638761 = weight(_text_:20 in 239) [ClassicSimilarity], result of:
          0.015638761 = score(doc=239,freq=6.0), product of:
            0.059459355 = queryWeight, product of:
              2.748821 = idf(docFreq=7692, maxDocs=44218)
              0.021630857 = queryNorm
            0.263016 = fieldWeight in 239, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.748821 = idf(docFreq=7692, maxDocs=44218)
              0.0390625 = fieldNorm(doc=239)
      0.033333335 = coord(1/30)
    
    Abstract
    A 5-year project to study scientific data uses in geography, starting in 1999, evolved into 20 years of research on data practices in sensor networks, environmental sciences, biology, seismology, undersea science, biomedicine, astronomy, and other fields. By emulating the "team science" approaches of the scientists studied, the UCLA Center for Knowledge Infrastructures accumulated a comprehensive collection of qualitative data about how scientists generate, manage, use, and reuse data across domains. Building upon Paul N. Edwards's model of "making global data"-collecting signals via consistent methods, technologies, and policies-to "make data global"-comparing and integrating those data, the research team has managed and exploited these data as a collaborative resource. This article reflects on the social, technical, organizational, economic, and policy challenges the team has encountered in creating new knowledge from data old and new. We reflect on continuity over generations of students and staff, transitions between grants, transfer of legacy data between software tools, research methods, and the role of professional data managers in the social sciences.
    Date
    20. 5.2021 18:35:59
  5. Lowe, D.B.; Dollinger, I.; Koster, T.; Herbert, B.E.: Text mining for type of research classification (2021) 0.00
    3.6116174E-4 = product of:
      0.010834851 = sum of:
        0.010834851 = weight(_text_:20 in 720) [ClassicSimilarity], result of:
          0.010834851 = score(doc=720,freq=2.0), product of:
            0.059459355 = queryWeight, product of:
              2.748821 = idf(docFreq=7692, maxDocs=44218)
              0.021630857 = queryNorm
            0.18222281 = fieldWeight in 720, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.748821 = idf(docFreq=7692, maxDocs=44218)
              0.046875 = fieldNorm(doc=720)
      0.033333335 = coord(1/30)
    
    Date
    16.12.2021 13:20:30