Search (15 results, page 1 of 1)

  • × theme_ss:"Data Mining"
  • × year_i:[2000 TO 2010}
  1. Maaten, L. van den: Learning a parametric embedding by preserving local structure (2009) 0.03
    0.027232561 = product of:
      0.054465123 = sum of:
        0.054465123 = product of:
          0.108930245 = sum of:
            0.108930245 = weight(_text_:t in 3883) [ClassicSimilarity], result of:
              0.108930245 = score(doc=3883,freq=8.0), product of:
                0.17876579 = queryWeight, product of:
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.04537884 = queryNorm
                0.60934615 = fieldWeight in 3883, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3883)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The paper presents a new unsupervised dimensionality reduction technique, called parametric t-SNE, that learns a parametric mapping between the high-dimensional data space and the low-dimensional latent space. Parametric t-SNE learns the parametric mapping in such a way that the local structure of the data is preserved as well as possible in the latent space. We evaluate the performance of parametric t-SNE in experiments on three datasets, in which we compare it to the performance of two other unsupervised parametric dimensionality reduction techniques. The results of experiments illustrate the strong performance of parametric t-SNE, in particular, in learning settings in which the dimensionality of the latent space is relatively low.
  2. Maaten, L. van den; Hinton, G.: Visualizing data using t-SNE (2008) 0.02
    0.02382353 = product of:
      0.04764706 = sum of:
        0.04764706 = product of:
          0.09529412 = sum of:
            0.09529412 = weight(_text_:t in 3888) [ClassicSimilarity], result of:
              0.09529412 = score(doc=3888,freq=12.0), product of:
                0.17876579 = queryWeight, product of:
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.04537884 = queryNorm
                0.53306687 = fieldWeight in 3888, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3888)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    We present a new technique called "t-SNE" that visualizes high-dimensional data by giving each datapoint a location in a two or three-dimensional map. The technique is a variation of Stochastic Neighbor Embedding (Hinton and Roweis, 2002) that is much easier to optimize, and produces significantly better visualizations by reducing the tendency to crowd points together in the center of the map. t-SNE is better than existing techniques at creating a single map that reveals structure at many different scales. This is particularly important for high-dimensional data that lie on several different, but related, low-dimensional manifolds, such as images of objects from multiple classes seen from multiple viewpoints. For visualizing the structure of very large data sets, we show how t-SNE can use random walks on neighborhood graphs to allow the implicit structure of all of the data to influence the way in which a subset of the data is displayed. We illustrate the performance of t-SNE on a wide variety of data sets and compare it with many other non-parametric visualization techniques, including Sammon mapping, Isomap, and Locally Linear Embedding. The visualizations produced by t-SNE are significantly better than those produced by the other techniques on almost all of the data sets.
  3. Witten, I.H.; Frank, E.: Data Mining : Praktische Werkzeuge und Techniken für das maschinelle Lernen (2000) 0.02
    0.023342196 = product of:
      0.04668439 = sum of:
        0.04668439 = product of:
          0.09336878 = sum of:
            0.09336878 = weight(_text_:t in 6833) [ClassicSimilarity], result of:
              0.09336878 = score(doc=6833,freq=2.0), product of:
                0.17876579 = queryWeight, product of:
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.04537884 = queryNorm
                0.5222967 = fieldWeight in 6833, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6833)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Footnote
    Rez. in: nfd 52(2001), H.7, S.427-428 (T. Mandl)
  4. Sperlich, T.: ¬Die Zukunft hat schon begonnen : Visualisierungssoftware in der praktischen Anwendung (2000) 0.02
    0.015561464 = product of:
      0.031122928 = sum of:
        0.031122928 = product of:
          0.062245857 = sum of:
            0.062245857 = weight(_text_:t in 5059) [ClassicSimilarity], result of:
              0.062245857 = score(doc=5059,freq=2.0), product of:
                0.17876579 = queryWeight, product of:
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.04537884 = queryNorm
                0.34819782 = fieldWeight in 5059, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5059)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  5. Brückner, T.; Dambeck, H.: Sortierautomaten : Grundlagen der Textklassifizierung (2003) 0.02
    0.015561464 = product of:
      0.031122928 = sum of:
        0.031122928 = product of:
          0.062245857 = sum of:
            0.062245857 = weight(_text_:t in 2398) [ClassicSimilarity], result of:
              0.062245857 = score(doc=2398,freq=2.0), product of:
                0.17876579 = queryWeight, product of:
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.04537884 = queryNorm
                0.34819782 = fieldWeight in 2398, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2398)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  6. Baumgartner, R.: Methoden und Werkzeuge zur Webdatenextraktion (2006) 0.01
    0.013616281 = product of:
      0.027232561 = sum of:
        0.027232561 = product of:
          0.054465123 = sum of:
            0.054465123 = weight(_text_:t in 5808) [ClassicSimilarity], result of:
              0.054465123 = score(doc=5808,freq=2.0), product of:
                0.17876579 = queryWeight, product of:
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.04537884 = queryNorm
                0.30467308 = fieldWeight in 5808, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5808)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Semantic Web: Wege zur vernetzten Wissensgesellschaft. Hrsg.: T. Pellegrini, u. A. Blumauer
  7. Kulathuramaiyer, N.; Maurer, H.: Implications of emerging data mining (2009) 0.01
    0.011671098 = product of:
      0.023342196 = sum of:
        0.023342196 = product of:
          0.04668439 = sum of:
            0.04668439 = weight(_text_:t in 3144) [ClassicSimilarity], result of:
              0.04668439 = score(doc=3144,freq=2.0), product of:
                0.17876579 = queryWeight, product of:
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.04537884 = queryNorm
                0.26114836 = fieldWeight in 3144, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3144)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Social Semantic Web: Web 2.0, was nun? Hrsg.: A. Blumauer u. T. Pellegrini
  8. Wu, T.; Pottenger, W.M.: ¬A semi-supervised active learning algorithm for information extraction from textual data (2005) 0.01
    0.009725915 = product of:
      0.01945183 = sum of:
        0.01945183 = product of:
          0.03890366 = sum of:
            0.03890366 = weight(_text_:t in 3237) [ClassicSimilarity], result of:
              0.03890366 = score(doc=3237,freq=2.0), product of:
                0.17876579 = queryWeight, product of:
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.04537884 = queryNorm
                0.21762364 = fieldWeight in 3237, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3237)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  9. Lackes, R.; Tillmanns, C.: Data Mining für die Unternehmenspraxis : Entscheidungshilfen und Fallstudien mit führenden Softwarelösungen (2006) 0.01
    0.00922231 = product of:
      0.01844462 = sum of:
        0.01844462 = product of:
          0.03688924 = sum of:
            0.03688924 = weight(_text_:22 in 1383) [ClassicSimilarity], result of:
              0.03688924 = score(doc=1383,freq=2.0), product of:
                0.15890898 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04537884 = queryNorm
                0.23214069 = fieldWeight in 1383, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1383)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 3.2008 14:46:06
  10. Heyer, G.; Quasthoff, U.; Wittig, T.: Text Mining : Wissensrohstoff Text. Konzepte, Algorithmen, Ergebnisse (2006) 0.01
    0.007780732 = product of:
      0.015561464 = sum of:
        0.015561464 = product of:
          0.031122928 = sum of:
            0.031122928 = weight(_text_:t in 5218) [ClassicSimilarity], result of:
              0.031122928 = score(doc=5218,freq=2.0), product of:
                0.17876579 = queryWeight, product of:
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.04537884 = queryNorm
                0.17409891 = fieldWeight in 5218, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9394085 = idf(docFreq=2338, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5218)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  11. Peters, G.; Gaese, V.: ¬Das DocCat-System in der Textdokumentation von G+J (2003) 0.01
    0.006148207 = product of:
      0.012296414 = sum of:
        0.012296414 = product of:
          0.024592828 = sum of:
            0.024592828 = weight(_text_:22 in 1507) [ClassicSimilarity], result of:
              0.024592828 = score(doc=1507,freq=2.0), product of:
                0.15890898 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04537884 = queryNorm
                0.15476047 = fieldWeight in 1507, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1507)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 4.2003 11:45:36
  12. Hölzig, C.: Google spürt Grippewellen auf : Die neue Anwendung ist bisher auf die USA beschränkt (2008) 0.01
    0.006148207 = product of:
      0.012296414 = sum of:
        0.012296414 = product of:
          0.024592828 = sum of:
            0.024592828 = weight(_text_:22 in 2403) [ClassicSimilarity], result of:
              0.024592828 = score(doc=2403,freq=2.0), product of:
                0.15890898 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04537884 = queryNorm
                0.15476047 = fieldWeight in 2403, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2403)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    3. 5.1997 8:44:22
  13. Lischka, K.: Spurensuche im Datenwust : Data-Mining-Software fahndet nach kriminellen Mitarbeitern, guten Kunden - und bald vielleicht auch nach Terroristen (2002) 0.00
    0.004611155 = product of:
      0.00922231 = sum of:
        0.00922231 = product of:
          0.01844462 = sum of:
            0.01844462 = weight(_text_:22 in 1178) [ClassicSimilarity], result of:
              0.01844462 = score(doc=1178,freq=2.0), product of:
                0.15890898 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04537884 = queryNorm
                0.116070345 = fieldWeight in 1178, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1178)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Content
    "Ob man als Terrorist einen Anschlag gegen die Vereinigten Staaten plant, als Kassierer Scheine aus der Kasse unterschlägt oder für bestimmte Produkte besonders gerne Geld ausgibt - einen Unterschied macht Data-Mining-Software da nicht. Solche Programme analysieren riesige Daten- mengen und fällen statistische Urteile. Mit diesen Methoden wollen nun die For- scher des "Information Awaren in den Vereinigten Staaten Spuren von Terroristen in den Datenbanken von Behörden und privaten Unternehmen wie Kreditkartenfirmen finden. 200 Millionen Dollar umfasst der Jahresetat für die verschiedenen Forschungsprojekte. Dass solche Software in der Praxis funktioniert, zeigen die steigenden Umsätze der Anbieter so genannter Customer-Relationship-Management-Software. Im vergangenen Jahr ist das Potenzial für analytische CRM-Anwendungen laut dem Marktforschungsinstitut IDC weltweit um 22 Prozent gewachsen, bis zum Jahr 2006 soll es in Deutschland mit einem jährlichen Plus von 14,1 Prozent so weitergehen. Und das trotz schwacher Konjunktur - oder gerade deswegen. Denn ähnlich wie Data-Mining der USRegierung helfen soll, Terroristen zu finden, entscheiden CRM-Programme heute, welche Kunden für eine Firma profitabel sind. Und welche es künftig sein werden, wie Manuela Schnaubelt, Sprecherin des CRM-Anbieters SAP, beschreibt: "Die Kundenbewertung ist ein zentraler Bestandteil des analytischen CRM. Sie ermöglicht es Unternehmen, sich auf die für sie wichtigen und richtigen Kunden zu fokussieren. Darüber hinaus können Firmen mit speziellen Scoring- Verfahren ermitteln, welche Kunden langfristig in welchem Maße zum Unternehmenserfolg beitragen." Die Folgen der Bewertungen sind für die Betroffenen nicht immer positiv: Attraktive Kunden profitieren von individuellen Sonderangeboten und besonderer Zuwendung. Andere hängen vielleicht so lauge in der Warteschleife des Telefonservice, bis die profitableren Kunden abgearbeitet sind. So könnte eine praktische Umsetzung dessen aussehen, was SAP-Spreche-rin Schnaubelt abstrakt beschreibt: "In vielen Unternehmen wird Kundenbewertung mit der klassischen ABC-Analyse durchgeführt, bei der Kunden anhand von Daten wie dem Umsatz kategorisiert werden. A-Kunden als besonders wichtige Kunden werden anders betreut als C-Kunden." Noch näher am geplanten Einsatz von Data-Mining zur Terroristenjagd ist eine Anwendung, die heute viele Firmen erfolgreich nutzen: Sie spüren betrügende Mitarbeiter auf. Werner Sülzer vom großen CRM-Anbieter NCR Teradata beschreibt die Möglichkeiten so: "Heute hinterlässt praktisch jeder Täter - ob Mitarbeiter, Kunde oder Lieferant - Datenspuren bei seinen wirtschaftskriminellen Handlungen. Es muss vorrangig darum gehen, einzelne Spuren zu Handlungsmustern und Täterprofilen zu verdichten. Das gelingt mittels zentraler Datenlager und hoch entwickelter Such- und Analyseinstrumente." Von konkreten Erfolgen sprich: Entlas-sungen krimineller Mitarbeiter-nach Einsatz solcher Programme erzählen Unternehmen nicht gerne. Matthias Wilke von der "Beratungsstelle für Technologiefolgen und Qualifizierung" (BTQ) der Gewerkschaft Verdi weiß von einem Fall 'aus der Schweiz. Dort setzt die Handelskette "Pick Pay" das Programm "Lord Lose Prevention" ein. Zwei Monate nach Einfüh-rung seien Unterschlagungen im Wert von etwa 200 000 Franken ermittelt worden. Das kostete mehr als 50 verdächtige Kassiererinnen und Kassierer den Job.
  14. Medien-Informationsmanagement : Archivarische, dokumentarische, betriebswirtschaftliche, rechtliche und Berufsbild-Aspekte ; [Frühjahrstagung der Fachgruppe 7 im Jahr 2000 in Weimar und Folgetagung 2001 in Köln] (2003) 0.00
    0.004611155 = product of:
      0.00922231 = sum of:
        0.00922231 = product of:
          0.01844462 = sum of:
            0.01844462 = weight(_text_:22 in 1833) [ClassicSimilarity], result of:
              0.01844462 = score(doc=1833,freq=2.0), product of:
                0.15890898 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04537884 = queryNorm
                0.116070345 = fieldWeight in 1833, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1833)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    11. 5.2008 19:49:22
  15. Information visualization in data mining and knowledge discovery (2002) 0.00
    0.0030741035 = product of:
      0.006148207 = sum of:
        0.006148207 = product of:
          0.012296414 = sum of:
            0.012296414 = weight(_text_:22 in 1789) [ClassicSimilarity], result of:
              0.012296414 = score(doc=1789,freq=2.0), product of:
                0.15890898 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04537884 = queryNorm
                0.07738023 = fieldWeight in 1789, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.015625 = fieldNorm(doc=1789)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    23. 3.2008 19:10:22

Languages

Types