Search (196 results, page 1 of 10)

  • × type_ss:"el"
  1. Lusti, M.: Data Warehousing and Data Mining : Eine Einführung in entscheidungsunterstützende Systeme (1999) 0.20
    0.19568293 = product of:
      0.39136586 = sum of:
        0.39136586 = sum of:
          0.3281657 = weight(_text_:mining in 4261) [ClassicSimilarity], result of:
            0.3281657 = score(doc=4261,freq=8.0), product of:
              0.3290036 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.058308665 = queryNorm
              0.9974533 = fieldWeight in 4261, product of:
                2.828427 = tf(freq=8.0), with freq of:
                  8.0 = termFreq=8.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.0625 = fieldNorm(doc=4261)
          0.06320015 = weight(_text_:22 in 4261) [ClassicSimilarity], result of:
            0.06320015 = score(doc=4261,freq=2.0), product of:
              0.204187 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.058308665 = queryNorm
              0.30952093 = fieldWeight in 4261, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=4261)
      0.5 = coord(1/2)
    
    Date
    17. 7.2002 19:22:06
    RSWK
    Data mining / Lehrbuch
    Subject
    Data mining / Lehrbuch
    Theme
    Data Mining
  2. Priss, U.: Description logic and faceted knowledge representation (1999) 0.09
    0.08523113 = product of:
      0.17046227 = sum of:
        0.17046227 = sum of:
          0.12306215 = weight(_text_:mining in 2655) [ClassicSimilarity], result of:
            0.12306215 = score(doc=2655,freq=2.0), product of:
              0.3290036 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.058308665 = queryNorm
              0.37404498 = fieldWeight in 2655, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.046875 = fieldNorm(doc=2655)
          0.047400113 = weight(_text_:22 in 2655) [ClassicSimilarity], result of:
            0.047400113 = score(doc=2655,freq=2.0), product of:
              0.204187 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.058308665 = queryNorm
              0.23214069 = fieldWeight in 2655, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2655)
      0.5 = coord(1/2)
    
    Abstract
    The term "facet" was introduced into the field of library classification systems by Ranganathan in the 1930's [Ranganathan, 1962]. A facet is a viewpoint or aspect. In contrast to traditional classification systems, faceted systems are modular in that a domain is analyzed in terms of baseline facets which are then synthesized. In this paper, the term "facet" is used in a broader meaning. Facets can describe different aspects on the same level of abstraction or the same aspect on different levels of abstraction. The notion of facets is related to database views, multicontexts and conceptual scaling in formal concept analysis [Ganter and Wille, 1999], polymorphism in object-oriented design, aspect-oriented programming, views and contexts in description logic and semantic networks. This paper presents a definition of facets in terms of faceted knowledge representation that incorporates the traditional narrower notion of facets and potentially facilitates translation between different knowledge representation formalisms. A goal of this approach is a modular, machine-aided knowledge base design mechanism. A possible application is faceted thesaurus construction for information retrieval and data mining. Reasoning complexity depends on the size of the modules (facets). A more general analysis of complexity will be left for future research.
    Date
    22. 1.2016 17:30:31
  3. Kleineberg, M.: Context analysis and context indexing : formal pragmatics in knowledge organization (2014) 0.08
    0.077174686 = product of:
      0.15434937 = sum of:
        0.15434937 = product of:
          0.4630481 = sum of:
            0.4630481 = weight(_text_:3a in 1826) [ClassicSimilarity], result of:
              0.4630481 = score(doc=1826,freq=2.0), product of:
                0.49434152 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.058308665 = queryNorm
                0.93669677 = fieldWeight in 1826, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1826)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Source
    http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=5&ved=0CDQQFjAE&url=http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F3131107&ei=HzFWVYvGMsiNsgGTyoFI&usg=AFQjCNE2FHUeR9oQTQlNC4TPedv4Mo3DaQ&sig2=Rlzpr7a3BLZZkqZCXXN_IA&bvm=bv.93564037,d.bGg&cad=rja
  4. Decimal Classification Editorial Policy Committee (2002) 0.07
    0.07230979 = sum of:
      0.04437901 = product of:
        0.13313703 = sum of:
          0.13313703 = weight(_text_:themes in 236) [ClassicSimilarity], result of:
            0.13313703 = score(doc=236,freq=2.0), product of:
              0.3748681 = queryWeight, product of:
                6.429029 = idf(docFreq=193, maxDocs=44218)
                0.058308665 = queryNorm
              0.35515702 = fieldWeight in 236, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                6.429029 = idf(docFreq=193, maxDocs=44218)
                0.0390625 = fieldNorm(doc=236)
        0.33333334 = coord(1/3)
      0.027930785 = product of:
        0.05586157 = sum of:
          0.05586157 = weight(_text_:22 in 236) [ClassicSimilarity], result of:
            0.05586157 = score(doc=236,freq=4.0), product of:
              0.204187 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.058308665 = queryNorm
              0.27358043 = fieldWeight in 236, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=236)
        0.5 = coord(1/2)
    
    Abstract
    The Decimal Classification Editorial Policy Committee (EPC) held its Meeting 117 at the Library Dec. 3-5, 2001, with chair Andrea Stamm (Northwestern University) presiding. Through its actions at this meeting, significant progress was made toward publication of DDC unabridged Edition 22 in mid-2003 and Abridged Edition 14 in early 2004. For Edition 22, the committee approved the revisions to two major segments of the classification: Table 2 through 55 Iran (the first half of the geographic area table) and 900 History and geography. EPC approved updates to several parts of the classification it had already considered: 004-006 Data processing, Computer science; 340 Law; 370 Education; 510 Mathematics; 610 Medicine; Table 3 issues concerning treatment of scientific and technical themes, with folklore, arts, and printing ramifications at 398.2 - 398.3, 704.94, and 758; Table 5 and Table 6 Ethnic Groups and Languages (portions concerning American native peoples and languages); and tourism issues at 647.9 and 790. Reports on the results of testing the approved 200 Religion and 305-306 Social groups schedules were received, as was a progress report on revision work for the manual being done by Ross Trotter (British Library, retired). Revisions for Abridged Edition 14 that received committee approval included 010 Bibliography; 070 Journalism; 150 Psychology; 370 Education; 380 Commerce, communications, and transportation; 621 Applied physics; 624 Civil engineering; and 629.8 Automatic control engineering. At the meeting the committee received print versions of _DC&_ numbers 4 and 5. Primarily for the use of Dewey translators, these cumulations list changes, substantive and cosmetic, to DDC Edition 21 and Abridged Edition 13 for the period October 1999 - December 2001. EPC will hold its Meeting 118 at the Library May 15-17, 2002.
  5. Winterhalter, C.: Licence to mine : ein Überblick über Rahmenbedingungen von Text and Data Mining und den aktuellen Stand der Diskussion (2016) 0.07
    0.071049966 = product of:
      0.14209993 = sum of:
        0.14209993 = product of:
          0.28419986 = sum of:
            0.28419986 = weight(_text_:mining in 673) [ClassicSimilarity], result of:
              0.28419986 = score(doc=673,freq=6.0), product of:
                0.3290036 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.058308665 = queryNorm
                0.86381996 = fieldWeight in 673, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.0625 = fieldNorm(doc=673)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Der Artikel gibt einen Überblick über die Möglichkeiten der Anwendung von Text and Data Mining (TDM) und ähnlichen Verfahren auf der Grundlage bestehender Regelungen in Lizenzverträgen zu kostenpflichtigen elektronischen Ressourcen, die Debatte über zusätzliche Lizenzen für TDM am Beispiel von Elseviers TDM Policy und den Stand der Diskussion über die Einführung von Schrankenregelungen im Urheberrecht für TDM zu nichtkommerziellen wissenschaftlichen Zwecken.
    Theme
    Data Mining
  6. Junger, U.; Schwens, U.: ¬Die inhaltliche Erschließung des schriftlichen kulturellen Erbes auf dem Weg in die Zukunft : Automatische Vergabe von Schlagwörtern in der Deutschen Nationalbibliothek (2017) 0.07
    0.07102594 = product of:
      0.14205188 = sum of:
        0.14205188 = sum of:
          0.10255179 = weight(_text_:mining in 3780) [ClassicSimilarity], result of:
            0.10255179 = score(doc=3780,freq=2.0), product of:
              0.3290036 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.058308665 = queryNorm
              0.31170416 = fieldWeight in 3780, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.0390625 = fieldNorm(doc=3780)
          0.039500095 = weight(_text_:22 in 3780) [ClassicSimilarity], result of:
            0.039500095 = score(doc=3780,freq=2.0), product of:
              0.204187 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.058308665 = queryNorm
              0.19345059 = fieldWeight in 3780, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=3780)
      0.5 = coord(1/2)
    
    Abstract
    Wir leben im 21. Jahrhundert, und vieles, was vor hundert und noch vor fünfzig Jahren als Science Fiction abgetan worden wäre, ist mittlerweile Realität. Raumsonden fliegen zum Mars, machen dort Experimente und liefern Daten zur Erde zurück. Roboter werden für Routineaufgaben eingesetzt, zum Beispiel in der Industrie oder in der Medizin. Digitalisierung, künstliche Intelligenz und automatisierte Verfahren sind kaum mehr aus unserem Alltag wegzudenken. Grundlage vieler Prozesse sind lernende Algorithmen. Die fortschreitende digitale Transformation ist global und umfasst alle Lebens- und Arbeitsbereiche: Wirtschaft, Gesellschaft und Politik. Sie eröffnet neue Möglichkeiten, von denen auch Bibliotheken profitieren. Der starke Anstieg digitaler Publikationen, die einen wichtigen und prozentual immer größer werdenden Teil des Kulturerbes darstellen, sollte für Bibliotheken Anlass sein, diese Möglichkeiten aktiv aufzugreifen und einzusetzen. Die Auswertbarkeit digitaler Inhalte, beispielsweise durch Text- and Data-Mining (TDM), und die Entwicklung technischer Verfahren, mittels derer Inhalte miteinander vernetzt und semantisch in Beziehung gesetzt werden können, bieten Raum, auch bibliothekarische Erschließungsverfahren neu zu denken. Daher beschäftigt sich die Deutsche Nationalbibliothek (DNB) seit einigen Jahren mit der Frage, wie sich die Prozesse bei der Erschließung von Medienwerken verbessern und maschinell unterstützen lassen. Sie steht dabei im regelmäßigen kollegialen Austausch mit anderen Bibliotheken, die sich ebenfalls aktiv mit dieser Fragestellung befassen, sowie mit europäischen Nationalbibliotheken, die ihrerseits Interesse an dem Thema und den Erfahrungen der DNB haben. Als Nationalbibliothek mit umfangreichen Beständen an digitalen Publikationen hat die DNB auch Expertise bei der digitalen Langzeitarchivierung aufgebaut und ist im Netzwerk ihrer Partner als kompetente Gesprächspartnerin geschätzt.
    Date
    19. 8.2017 9:24:22
  7. Perovsek, M.; Kranjca, J.; Erjaveca, T.; Cestnika, B.; Lavraca, N.: TextFlows : a visual programming platform for text mining and natural language processing (2016) 0.07
    0.06879383 = product of:
      0.13758767 = sum of:
        0.13758767 = product of:
          0.27517533 = sum of:
            0.27517533 = weight(_text_:mining in 2697) [ClassicSimilarity], result of:
              0.27517533 = score(doc=2697,freq=10.0), product of:
                0.3290036 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.058308665 = queryNorm
                0.83639 = fieldWeight in 2697, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2697)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Text mining and natural language processing are fast growing areas of research, with numerous applications in business, science and creative industries. This paper presents TextFlows, a web-based text mining and natural language processing platform supporting workflow construction, sharing and execution. The platform enables visual construction of text mining workflows through a web browser, and the execution of the constructed workflows on a processing cloud. This makes TextFlows an adaptable infrastructure for the construction and sharing of text processing workflows, which can be reused in various applications. The paper presents the implemented text mining and language processing modules, and describes some precomposed workflows. Their features are demonstrated on three use cases: comparison of document classifiers and of different part-of-speech taggers on a text categorization problem, and outlier detection in document corpora.
  8. Dowding, H.; Gengenbach, M.; Graham, B.; Meister, S.; Moran, J.; Peltzman, S.; Seifert, J.; Waugh, D.: OSS4EVA: using open-source tools to fulfill digital preservation requirements (2016) 0.06
    0.064129055 = sum of:
      0.04437901 = product of:
        0.13313703 = sum of:
          0.13313703 = weight(_text_:themes in 3200) [ClassicSimilarity], result of:
            0.13313703 = score(doc=3200,freq=2.0), product of:
              0.3748681 = queryWeight, product of:
                6.429029 = idf(docFreq=193, maxDocs=44218)
                0.058308665 = queryNorm
              0.35515702 = fieldWeight in 3200, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                6.429029 = idf(docFreq=193, maxDocs=44218)
                0.0390625 = fieldNorm(doc=3200)
        0.33333334 = coord(1/3)
      0.019750047 = product of:
        0.039500095 = sum of:
          0.039500095 = weight(_text_:22 in 3200) [ClassicSimilarity], result of:
            0.039500095 = score(doc=3200,freq=2.0), product of:
              0.204187 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.058308665 = queryNorm
              0.19345059 = fieldWeight in 3200, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=3200)
        0.5 = coord(1/2)
    
    Abstract
    This paper builds on the findings of a workshop held at the 2015 International Conference on Digital Preservation (iPRES), entitled, "Using Open-Source Tools to Fulfill Digital Preservation Requirements" (OSS4PRES hereafter). This day-long workshop brought together participants from across the library and archives community, including practitioners proprietary vendors, and representatives from open-source projects. The resulting conversations were surprisingly revealing: while OSS' significance within the preservation landscape was made clear, participants noted that there are a number of roadblocks that discourage or altogether prevent its use in many organizations. Overcoming these challenges will be necessary to further widespread, sustainable OSS adoption within the digital preservation community. This article will mine the rich discussions that took place at OSS4PRES to (1) summarize the workshop's key themes and major points of debate, (2) provide a comprehensive analysis of the opportunities, gaps, and challenges that using OSS entails at a philosophical, institutional, and individual level, and (3) offer a tangible set of recommendations for future work designed to broaden community engagement and enhance the sustainability of open source initiatives, drawing on both participants' experience as well as additional research.
    Date
    28.10.2016 18:22:33
  9. Popper, K.R.: Three worlds : the Tanner lecture on human values. Deliverd at the University of Michigan, April 7, 1978 (1978) 0.06
    0.06173975 = product of:
      0.1234795 = sum of:
        0.1234795 = product of:
          0.3704385 = sum of:
            0.3704385 = weight(_text_:3a in 230) [ClassicSimilarity], result of:
              0.3704385 = score(doc=230,freq=2.0), product of:
                0.49434152 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.058308665 = queryNorm
                0.7493574 = fieldWeight in 230, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0625 = fieldNorm(doc=230)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Source
    https%3A%2F%2Ftannerlectures.utah.edu%2F_documents%2Fa-to-z%2Fp%2Fpopper80.pdf&usg=AOvVaw3f4QRTEH-OEBmoYr2J_c7H
  10. Bradford, R.B.: Relationship discovery in large text collections using Latent Semantic Indexing (2006) 0.06
    0.05682075 = product of:
      0.1136415 = sum of:
        0.1136415 = sum of:
          0.08204143 = weight(_text_:mining in 1163) [ClassicSimilarity], result of:
            0.08204143 = score(doc=1163,freq=2.0), product of:
              0.3290036 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.058308665 = queryNorm
              0.24936332 = fieldWeight in 1163, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.03125 = fieldNorm(doc=1163)
          0.031600077 = weight(_text_:22 in 1163) [ClassicSimilarity], result of:
            0.031600077 = score(doc=1163,freq=2.0), product of:
              0.204187 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.058308665 = queryNorm
              0.15476047 = fieldWeight in 1163, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=1163)
      0.5 = coord(1/2)
    
    Source
    Proceedings of the Fourth Workshop on Link Analysis, Counterterrorism, and Security, SIAM Data Mining Conference, Bethesda, MD, 20-22 April, 2006. [http://www.siam.org/meetings/sdm06/workproceed/Link%20Analysis/15.pdf]
  11. Hensinger, P.: Trojanisches Pferd "Digitale Bildung" : Auf dem Weg zur Konditionierungsanstalt in einer Schule ohne Lehrer? (2017) 0.06
    0.05682075 = product of:
      0.1136415 = sum of:
        0.1136415 = sum of:
          0.08204143 = weight(_text_:mining in 5000) [ClassicSimilarity], result of:
            0.08204143 = score(doc=5000,freq=2.0), product of:
              0.3290036 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.058308665 = queryNorm
              0.24936332 = fieldWeight in 5000, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.03125 = fieldNorm(doc=5000)
          0.031600077 = weight(_text_:22 in 5000) [ClassicSimilarity], result of:
            0.031600077 = score(doc=5000,freq=2.0), product of:
              0.204187 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.058308665 = queryNorm
              0.15476047 = fieldWeight in 5000, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=5000)
      0.5 = coord(1/2)
    
    Abstract
    Wir hatten schon viele Schulreformen, und nun wird von der Kultusministerkonferenz eine weitere angekün-digt, die "Digitale Bildung": Unterricht mit digitalen Medien wie Smartphone und Tablet-PC über WLAN.1Medien und Bildungspolitiker predigen Eltern, ihre Kinder seien in Schule und Beruf chancenlos, wenn sie nicht schon in der Grundschule Apps programmieren lernen.Die Hauptinitiative der Digitalisierung der Bildung kommt von der IT-Branche. Im Zwischenbericht der Platt-form "Digitalisierung in Bildung und Wissenschaft" steht, wer das Bundeswissenschaftsministerium berät - nämlich Akteure der IT-Wirtschaft: Vom Bitkom, der Gesellschaft für Informatik (GI) über Microsoft, SAP bis zur Telekom sind alle vertreten (BUNDESMINISTERIUM 2016:23). Nicht vertreten dagegen sind Kinderärzte, Päda-gogen, Lernpsychologen oder Neurowissenschaftler, die sich mit den Folgen der Nutzung von Bildschirm-medien bei Kindern und Jugendlichen beschäftigen. Die New York Times schlägt in einer Analyse Alarm: "How Google Took Over the Classroom" (13.05.2017).2 Mit ausgeklügelten Methoden, den Hype um digitale Medien nutzend, greift Google nach der Kontrolle des US-Bildungswesens, auch der Kontrolle über die Inhalte. Wer bei der Analyse und Bewertung dieser Entwicklung nur fragt "Nützen digitale Medien im Unterricht?", verengt den Blick, reduziert auf Methodik und Didaktik und schließt Gesamtzusammenhänge aus. Denn die digitalen Medien sind mehr als nur Unterrichts-Hilfsmittel. Diesen Tunnelblick weitet die IT-Unternehmerin Yvonne Hofstetter. Sie schreibt in ihrem Buch "Das Ende der Demokratie": "Mit der Digitalisierung verwandeln wir unser Leben, privat wie beruflich, in einen Riesencomputer. Alles wird gemessen, gespeichert, analysiert und prognostiziert, um es anschließend zu steuern und zu optimieren"(HOFSTETTER 2016:37). Grundlage dafür ist das Data-Mining - das Sammeln von Daten - für BigData Analysen. Die Haupt-Schürfwerkzeuge dazu sind dasSmartphone, der TabletPC und das WLAN-Netz.
    Date
    22. 2.2019 11:45:19
  12. Jäger, L.: Von Big Data zu Big Brother (2018) 0.06
    0.05682075 = product of:
      0.1136415 = sum of:
        0.1136415 = sum of:
          0.08204143 = weight(_text_:mining in 5234) [ClassicSimilarity], result of:
            0.08204143 = score(doc=5234,freq=2.0), product of:
              0.3290036 = queryWeight, product of:
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.058308665 = queryNorm
              0.24936332 = fieldWeight in 5234, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.642448 = idf(docFreq=425, maxDocs=44218)
                0.03125 = fieldNorm(doc=5234)
          0.031600077 = weight(_text_:22 in 5234) [ClassicSimilarity], result of:
            0.031600077 = score(doc=5234,freq=2.0), product of:
              0.204187 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.058308665 = queryNorm
              0.15476047 = fieldWeight in 5234, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=5234)
      0.5 = coord(1/2)
    
    Date
    22. 1.2018 11:33:49
    Theme
    Data Mining
  13. Blosser, J.; Michaelson, R.; Routh. R.; Xia, P.: Defining the landscape of Web resources : Concluding Report of the BAER Web Resources Sub-Group (2000) 0.05
    0.051303253 = sum of:
      0.035503212 = product of:
        0.10650963 = sum of:
          0.10650963 = weight(_text_:themes in 1447) [ClassicSimilarity], result of:
            0.10650963 = score(doc=1447,freq=2.0), product of:
              0.3748681 = queryWeight, product of:
                6.429029 = idf(docFreq=193, maxDocs=44218)
                0.058308665 = queryNorm
              0.28412563 = fieldWeight in 1447, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                6.429029 = idf(docFreq=193, maxDocs=44218)
                0.03125 = fieldNorm(doc=1447)
        0.33333334 = coord(1/3)
      0.015800038 = product of:
        0.031600077 = sum of:
          0.031600077 = weight(_text_:22 in 1447) [ClassicSimilarity], result of:
            0.031600077 = score(doc=1447,freq=2.0), product of:
              0.204187 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.058308665 = queryNorm
              0.15476047 = fieldWeight in 1447, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=1447)
        0.5 = coord(1/2)
    
    Abstract
    The BAER Web Resources Group was charged in October 1999 with defining and describing the parameters of electronic resources that do not clearly belong to the categories being defined by the BAER Digital Group or the BAER Electronic Journals Group. After some difficulty identifying precisely which resources fell under the Group's charge, we finally named the following types of resources for our consideration: web sites, electronic texts, indexes, databases and abstracts, online reference resources, and networked and non-networked CD-ROMs. Electronic resources are a vast and growing collection that touch nearly every department within the Library. It is unrealistic to think one department can effectively administer all aspects of the collection. The Group then began to focus on the concern of bibliographic access to these varied resources, and to define parameters for handling or processing them within the Library. Some key elements became evident as the work progressed. * Selection process of resources to be acquired for the collection * Duplication of effort * Use of CORC * Resource Finder design * Maintenance of Resource Finder * CD-ROMs not networked * Communications * Voyager search limitations. An unexpected collaboration with the Web Development Committee on the Resource Finder helped to steer the Group to more detailed descriptions of bibliographic access. This collaboration included development of data elements for the Resource Finder database, and some discussions on Library staff processing of the resources. The Web Resources Group invited expert testimony to help the Group broaden its view to envision public use of the resources and discuss concerns related to technical services processing. The first testimony came from members of the Resource Finder Committee. Some background information on the Web Development Resource Finder Committee was shared. The second testimony was from librarians who select electronic texts. Three main themes were addressed: accessing CD-ROMs; the issue of including non-networked CD-ROMs in the Resource Finder; and, some special concerns about electronic texts. The third testimony came from librarians who select indexes and abstracts and also provide Reference services. Appendices to this report include minutes of the meetings with the experts (Appendix A), a list of proposed data elements to be used in the Resource Finder (Appendix B), and recommendations made to the Resource Finder Committee (Appendix C). Below are summaries of the key elements.
    Date
    21. 4.2002 10:22:31
  14. Mohr, J.W.; Bogdanov, P.: Topic models : what they are and why they matter (2013) 0.04
    0.04350904 = product of:
      0.08701808 = sum of:
        0.08701808 = product of:
          0.17403616 = sum of:
            0.17403616 = weight(_text_:mining in 1142) [ClassicSimilarity], result of:
              0.17403616 = score(doc=1142,freq=4.0), product of:
                0.3290036 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.058308665 = queryNorm
                0.5289795 = fieldWeight in 1142, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1142)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    We provide a brief, non-technical introduction to the text mining methodology known as "topic modeling." We summarize the theory and background of the method and discuss what kinds of things are found by topic models. Using a text corpus comprised of the eight articles from the special issue of Poetics on the subject of topic models, we run a topic model on these articles, both as a way to introduce the methodology and also to help summarize some of the ways in which social and cultural scientists are using topic models. We review some of the critiques and debates over the use of the method and finally, we link these developments back to some of the original innovations in the field of content analysis that were pioneered by Harold D. Lasswell and colleagues during and just after World War II.
    Theme
    Data Mining
  15. Eckert, K: ¬The ICE-map visualization (2011) 0.04
    0.041020714 = product of:
      0.08204143 = sum of:
        0.08204143 = product of:
          0.16408285 = sum of:
            0.16408285 = weight(_text_:mining in 4743) [ClassicSimilarity], result of:
              0.16408285 = score(doc=4743,freq=2.0), product of:
                0.3290036 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.058308665 = queryNorm
                0.49872664 = fieldWeight in 4743, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4743)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    In this paper, we describe in detail the Information Content Evaluation Map (ICE-Map Visualization, formerly referred to as IC Difference Analysis). The ICE-Map Visualization is a visual data mining approach for all kinds of concept hierarchies that uses statistics about the concept usage to help a user in the evaluation and maintenance of the hierarchy. It consists of a statistical framework that employs the the notion of information content from information theory, as well as a visualization of the hierarchy and the result of the statistical analysis by means of a treemap.
  16. Bauckhage, C.: Moderne Textanalyse : neues Wissen für intelligente Lösungen (2016) 0.04
    0.041020714 = product of:
      0.08204143 = sum of:
        0.08204143 = product of:
          0.16408285 = sum of:
            0.16408285 = weight(_text_:mining in 2568) [ClassicSimilarity], result of:
              0.16408285 = score(doc=2568,freq=2.0), product of:
                0.3290036 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.058308665 = queryNorm
                0.49872664 = fieldWeight in 2568, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2568)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Theme
    Data Mining
  17. Wattenberg, M.; Viégas, F.; Johnson, I.: How to use t-SNE effectively (2016) 0.04
    0.041020714 = product of:
      0.08204143 = sum of:
        0.08204143 = product of:
          0.16408285 = sum of:
            0.16408285 = weight(_text_:mining in 3887) [ClassicSimilarity], result of:
              0.16408285 = score(doc=3887,freq=2.0), product of:
                0.3290036 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.058308665 = queryNorm
                0.49872664 = fieldWeight in 3887, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3887)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Theme
    Data Mining
  18. Nohr, H.: Big Data im Lichte der EU-Datenschutz-Grundverordnung (2017) 0.04
    0.041020714 = product of:
      0.08204143 = sum of:
        0.08204143 = product of:
          0.16408285 = sum of:
            0.16408285 = weight(_text_:mining in 4076) [ClassicSimilarity], result of:
              0.16408285 = score(doc=4076,freq=2.0), product of:
                0.3290036 = queryWeight, product of:
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.058308665 = queryNorm
                0.49872664 = fieldWeight in 4076, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.642448 = idf(docFreq=425, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4076)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Theme
    Data Mining
  19. Shala, E.: ¬Die Autonomie des Menschen und der Maschine : gegenwärtige Definitionen von Autonomie zwischen philosophischem Hintergrund und technologischer Umsetzbarkeit (2014) 0.04
    0.038587343 = product of:
      0.077174686 = sum of:
        0.077174686 = product of:
          0.23152405 = sum of:
            0.23152405 = weight(_text_:3a in 4388) [ClassicSimilarity], result of:
              0.23152405 = score(doc=4388,freq=2.0), product of:
                0.49434152 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.058308665 = queryNorm
                0.46834838 = fieldWeight in 4388, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4388)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Footnote
    Vgl. unter: https://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=2&cad=rja&uact=8&ved=2ahUKEwizweHljdbcAhVS16QKHXcFD9QQFjABegQICRAB&url=https%3A%2F%2Fwww.researchgate.net%2Fpublication%2F271200105_Die_Autonomie_des_Menschen_und_der_Maschine_-_gegenwartige_Definitionen_von_Autonomie_zwischen_philosophischem_Hintergrund_und_technologischer_Umsetzbarkeit_Redigierte_Version_der_Magisterarbeit_Karls&usg=AOvVaw06orrdJmFF2xbCCp_hL26q.
  20. Dietz, K.: en.wikipedia.org > 6 Mio. Artikel (2020) 0.04
    0.038587343 = product of:
      0.077174686 = sum of:
        0.077174686 = product of:
          0.23152405 = sum of:
            0.23152405 = weight(_text_:3a in 5669) [ClassicSimilarity], result of:
              0.23152405 = score(doc=5669,freq=2.0), product of:
                0.49434152 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.058308665 = queryNorm
                0.46834838 = fieldWeight in 5669, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5669)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Content
    "Die Englischsprachige Wikipedia verfügt jetzt über mehr als 6 Millionen Artikel. An zweiter Stelle kommt die deutschsprachige Wikipedia mit 2.3 Millionen Artikeln, an dritter Stelle steht die französischsprachige Wikipedia mit 2.1 Millionen Artikeln (via Researchbuzz: Firehose <https://rbfirehose.com/2020/01/24/techcrunch-wikipedia-now-has-more-than-6-million-articles-in-english/> und Techcrunch <https://techcrunch.com/2020/01/23/wikipedia-english-six-million-articles/?utm_source=feedburner&utm_medium=feed&utm_campaign=Feed%3A+Techcrunch+%28TechCrunch%29&guccounter=1&guce_referrer=aHR0cHM6Ly9yYmZpcmVob3NlLmNvbS8yMDIwLzAxLzI0L3RlY2hjcnVuY2gtd2lraXBlZGlhLW5vdy1oYXMtbW9yZS10aGFuLTYtbWlsbGlvbi1hcnRpY2xlcy1pbi1lbmdsaXNoLw&guce_referrer_sig=AQAAAK0zHfjdDZ_spFZBF_z-zDjtL5iWvuKDumFTzm4HvQzkUfE2pLXQzGS6FGB_y-VISdMEsUSvkNsg2U_NWQ4lwWSvOo3jvXo1I3GtgHpP8exukVxYAnn5mJspqX50VHIWFADHhs5AerkRn3hMRtf_R3F1qmEbo8EROZXp328HMC-o>). 250120 via digithek ch = #fineBlog s.a.: Angesichts der Veröffentlichung des 6-millionsten Artikels vergangene Woche in der englischsprachigen Wikipedia hat die Community-Zeitungsseite "Wikipedia Signpost" ein Moratorium bei der Veröffentlichung von Unternehmensartikeln gefordert. Das sei kein Vorwurf gegen die Wikimedia Foundation, aber die derzeitigen Maßnahmen, um die Enzyklopädie gegen missbräuchliches undeklariertes Paid Editing zu schützen, funktionierten ganz klar nicht. *"Da die ehrenamtlichen Autoren derzeit von Werbung in Gestalt von Wikipedia-Artikeln überwältigt werden, und da die WMF nicht in der Lage zu sein scheint, dem irgendetwas entgegenzusetzen, wäre der einzige gangbare Weg für die Autoren, fürs erste die Neuanlage von Artikeln über Unternehmen zu untersagen"*, schreibt der Benutzer Smallbones in seinem Editorial <https://en.wikipedia.org/wiki/Wikipedia:Wikipedia_Signpost/2020-01-27/From_the_editor> zur heutigen Ausgabe."

Years

Languages

  • e 97
  • d 93
  • el 2
  • a 1
  • nl 1
  • More… Less…

Types

  • a 95
  • i 10
  • m 5
  • r 3
  • b 2
  • s 2
  • n 1
  • p 1
  • x 1
  • More… Less…