Search (180 results, page 1 of 9)

  • × theme_ss:"Visualisierung"
  1. Ohly, H.P.: Erstellung und Interpretation von semantischen Karten am Beispiel des Themas 'Soziologische Beratung' (2004) 0.04
    0.039103523 = product of:
      0.078207046 = sum of:
        0.07601224 = weight(_text_:da in 3176) [ClassicSimilarity], result of:
          0.07601224 = score(doc=3176,freq=2.0), product of:
            0.20483522 = queryWeight, product of:
              4.7981725 = idf(docFreq=990, maxDocs=44218)
              0.04269026 = queryNorm
            0.3710897 = fieldWeight in 3176, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.7981725 = idf(docFreq=990, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3176)
        0.0021948046 = product of:
          0.004389609 = sum of:
            0.004389609 = weight(_text_:a in 3176) [ClassicSimilarity], result of:
              0.004389609 = score(doc=3176,freq=2.0), product of:
                0.049223874 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04269026 = queryNorm
                0.089176424 = fieldWeight in 3176, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3176)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Bei der Analyse von Informationsströmen und -systemen mit statistischen Methoden werden die Ergebnisse gerne in Grafiken dargestellt, da diese intuitiv schneller zu erfassen sind und auch Laien ohne tiefere statistische Vorbildung eine Anschauung bekommen können. Klassisches Beispiel ist etwa die graphische Darstellung der Verluste des napoleonischen Heeres in Russland (Abb. 1). Unbeachtet bleibt dabei oft, dass trotz Einfachheit der Darstellung meist große Mengen von Daten herangezogen werden und diese dann lediglich nach wenigen Gesichtspunkten in eine Grafik projiziert werdens, was leicht auch zu Fehleinschätzungen führen kann. Es sind darum geeignete Verfahren auszuwählen, die eine adäquate und möglichst 'objektive' Interpretation ermöglichen.
    Type
    a
  2. Passath, C.: Information-Panels : Die Informationsvermittler der Zukunft (2005) 0.03
    0.033517305 = product of:
      0.06703461 = sum of:
        0.065153345 = weight(_text_:da in 4991) [ClassicSimilarity], result of:
          0.065153345 = score(doc=4991,freq=2.0), product of:
            0.20483522 = queryWeight, product of:
              4.7981725 = idf(docFreq=990, maxDocs=44218)
              0.04269026 = queryNorm
            0.31807688 = fieldWeight in 4991, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.7981725 = idf(docFreq=990, maxDocs=44218)
              0.046875 = fieldNorm(doc=4991)
        0.0018812611 = product of:
          0.0037625222 = sum of:
            0.0037625222 = weight(_text_:a in 4991) [ClassicSimilarity], result of:
              0.0037625222 = score(doc=4991,freq=2.0), product of:
                0.049223874 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.04269026 = queryNorm
                0.07643694 = fieldWeight in 4991, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4991)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Je mehr Informationen gleichzeitig, übersichtlich dargestellt und überwacht werden können, desto höher wird der Informations-Nutzen für einen Besucher oder potentiellen Kunden von ihnen sein. Da wir in einer Zeit der Informationsüberflutung leben, hilft uns das Information-Panel in Zukunft als multimediales Informationssystem. Ein Interface ist in unserem heutigen multimedialen Umfeld ein Bestandteil eines Systems, das dem Austausch von Informationen dient. Durch Information-Panels (sog. I-Panels) kann der Mensch mit Geräten interagieren, indem er sich wahlweise die für ihn masßgeschneiderten Informationen darstellen lässt. Für den Interface-Theoretiker Artur P. Schmidt können Information-Panels heute als eine Art Enzyklopädie für Informationen und Nachrichten aller Art dienen, wie sein Internet-Projekt "Der Wissensnavigator" belegt. Das Vorbild für multimediale Panels ist die geordnete Verbindung von Inhalten. Das Information-Panel als MenschMaschine-Interface kann zum "Punkt der Begegnung" oder "Kopplung zwischen zwei oder mehr Systemen" werden. Es übernimmt eine Übersetzung- und Vermittlungsfunktion.
    Type
    a
  3. Waechter, U.: Visualisierung von Netzwerkstrukturen (2002) 0.02
    0.021717783 = product of:
      0.08687113 = sum of:
        0.08687113 = weight(_text_:da in 1735) [ClassicSimilarity], result of:
          0.08687113 = score(doc=1735,freq=2.0), product of:
            0.20483522 = queryWeight, product of:
              4.7981725 = idf(docFreq=990, maxDocs=44218)
              0.04269026 = queryNorm
            0.42410251 = fieldWeight in 1735, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.7981725 = idf(docFreq=990, maxDocs=44218)
              0.0625 = fieldNorm(doc=1735)
      0.25 = coord(1/4)
    
    Abstract
    Das WWW entwickelte sich aus dem Bedürfnis, textuelle Information einfach und schnell zu durchforsten. Dabei entstand das Konzept des 'Hyperlinks', womit es möglich ist, Texte miteinander zu verknüpfen. Die Anzahl der Webseiten nahm mit der Verbreitung des WWW rapide zu. Das Problem heutzutage ist: Es gibt prinzipiell jede Art von Information im Internet, doch wie kommt man da dran?
  4. Platis, N. et al.: Visualization of uncertainty in tag clouds (2016) 0.02
    0.016027568 = product of:
      0.06411027 = sum of:
        0.06411027 = sum of:
          0.0062708696 = weight(_text_:a in 2755) [ClassicSimilarity], result of:
            0.0062708696 = score(doc=2755,freq=2.0), product of:
              0.049223874 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.04269026 = queryNorm
              0.12739488 = fieldWeight in 2755, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.078125 = fieldNorm(doc=2755)
          0.057839405 = weight(_text_:22 in 2755) [ClassicSimilarity], result of:
            0.057839405 = score(doc=2755,freq=2.0), product of:
              0.149494 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04269026 = queryNorm
              0.38690117 = fieldWeight in 2755, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.078125 = fieldNorm(doc=2755)
      0.25 = coord(1/4)
    
    Date
    1. 2.2016 18:25:22
    Type
    a
  5. Börner, K.: Atlas of knowledge : anyone can map (2015) 0.01
    0.013210221 = product of:
      0.052840885 = sum of:
        0.052840885 = sum of:
          0.0037625222 = weight(_text_:a in 3355) [ClassicSimilarity], result of:
            0.0037625222 = score(doc=3355,freq=2.0), product of:
              0.049223874 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.04269026 = queryNorm
              0.07643694 = fieldWeight in 3355, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=3355)
          0.049078364 = weight(_text_:22 in 3355) [ClassicSimilarity], result of:
            0.049078364 = score(doc=3355,freq=4.0), product of:
              0.149494 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04269026 = queryNorm
              0.32829654 = fieldWeight in 3355, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=3355)
      0.25 = coord(1/4)
    
    Content
    One of a series of three publications influenced by the travelling exhibit Places & Spaces: Mapping Science, curated by the Cyberinfrastructure for Network Science Center at Indiana University. - Additional materials can be found at http://http://scimaps.org/atlas2. Erweitert durch: Börner, Katy. Atlas of Science: Visualizing What We Know.
    Date
    22. 1.2017 16:54:03
    22. 1.2017 17:10:56
  6. Teutsch, K.: ¬Die Welt ist doch eine Scheibe : Google-Herausforderer eyePlorer (2009) 0.01
    0.011755096 = product of:
      0.047020383 = sum of:
        0.047020383 = weight(_text_:da in 2678) [ClassicSimilarity], result of:
          0.047020383 = score(doc=2678,freq=6.0), product of:
            0.20483522 = queryWeight, product of:
              4.7981725 = idf(docFreq=990, maxDocs=44218)
              0.04269026 = queryNorm
            0.22955224 = fieldWeight in 2678, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.7981725 = idf(docFreq=990, maxDocs=44218)
              0.01953125 = fieldNorm(doc=2678)
      0.25 = coord(1/4)
    
    Content
    Eine neue visuelle Ordnung Martin Hirsch ist der Enkel des Nobelpreisträgers Werner Heisenberg. Außerdem ist er Hirnforscher und beschäftigt sich seit Jahren mit der Frage: Was tut mein Kopf eigentlich, während ich hirnforsche? Ralf von Grafenstein ist Marketingexperte und spezialisiert auf Dienstleistungen im Internet. Zusammen haben sie also am 1. Dezember 2008 eine Firma in Berlin gegründet, deren Heiliger Gral besagte Scheibe ist, auf der - das ist die Idee - bald die ganze Welt, die Internetwelt zumindest, Platz finden soll. Die Scheibe heißt eyePlorer, was sich als Aufforderung an ihre Nutzer versteht. Die sollen auf einer neuartigen, eben scheibenförmigen Plattform die unermesslichen Datensätze des Internets in eine neue visuelle Ordnung bringen. Der Schlüssel dafür, da waren sich Hirsch und von Grafenstein sicher, liegt in der Hirnforschung, denn warum nicht die assoziativen Fähigkeiten des Menschen auf Suchmaschinen übertragen? Anbieter wie Google lassen von solchen Ansätzen bislang die Finger. Hier setzt man dafür auf Volltextprogramme, also sprachbegabte Systeme, die letztlich aber, genau wie die Schlagwortsuche, nur zu opak gerankten Linksammlungen führen. Weiter als auf Seite zwei des Suchergebnisses wagt sich der träge Nutzer meistens nicht vor. Weil sie niemals wahrgenommen wird, fällt eine Menge möglicherweise kostbare Information unter den Tisch.
    Einstein, Weizsäcker und Hitler Zu Demonstrationszwecken wird die eyePlorer-Scheibe an die Wand projiziert. Gibt man im kleinen Suchfeld in der Mitte den Namen Werner Heisenberg ein, verwandelt sich die Scheibe in einen Tortenboden. Die einzelnen Stücke entsprechen Kategorien wie "Person", "Technologie" oder "Organisation". Sie selbst sind mit bunten Knöpfen bedeckt, unter denen sich die Informationen verbergen. So kommt es, dass man beim Thema Heisenberg nicht nur auf die Kollegen Einstein, Weizsäcker und Schrödinger trifft, sondern auch auf Adolf Hitler. Ein Klick auf den entsprechenden Button stellt unter anderem heraus: Heisenberg kam 1933 unter Beschuss der SS, weil er sich nicht vor den Karren einer antisemitischen Physikbewegung spannen ließ. Nach diesem Prinzip spült die frei assoziierende Maschine vollautomatisch immer wieder neue Fakten an, um die der Nutzer zwar nicht gebeten hat, die ihn bei seiner Recherche aber möglicherweise unterstützen und die er später - die Maschine ist noch ausbaubedürftig - auch modellieren darf. Aber will man das, sich von einer Maschine beraten lassen? "Google ist wie ein Zoo", sekundiert Ralf von Grafenstein. "In einem Gehege steht eine Giraffe, im anderen ein Raubtier, aber die sind klar getrennt voneinander durch Gitter und Wege. Es gibt keine Möglichkeit, sie zusammen anzuschauen. Da kommen wir ins Spiel. Wir können Äpfel mit Birnen vergleichen!" Die Welt ist eine Scheibe oder die Scheibe eben eine Welt, auf der vieles mit vielem zusammenhängt und manches auch mit nichts. Der Vorteil dieser Maschine ist, dass sie in Zukunft Sinn stiften könnte, wo andere nur spröde auf Quellen verweisen. "Google ist ja ein unheimlich heterogenes Erlebnis mit ständigen Wartezeiten und Mausklicks dazwischen. Das kostet mich viel zu viel Metagedankenkraft", sagt Hirsch. "Wir wollten eine Maschine mit einer ästhetisch ansprechenden Umgebung bauen, aus der ich mich kaum wegbewege, denn sie liefert mir Informationen in meinen Gedanken hinein."
    Wenn die Maschine denkt Zur Hybris des Projekts passt, dass der eyePlorer ursprünglich HAL heißen sollte - wie der außer Rand und Band geratene Bordcomputer aus Kubricks "2001: Odyssee im Weltraum". Wenn man die Buchstaben aber jeweils um eine Alphabetposition nach rechts verrückt, ergibt sich IBM. Was passiert mit unserem Wissen, wenn die Maschine selbst anfängt zu denken? Ralf von Grafenstein macht ein ernstes Gesicht. "Es ist nicht unser Ansinnen, sie alleinzulassen. Es geht bei uns ja nicht nur darum, zu finden, sondern auch mitzumachen. Die Community ist wichtig. Der Dialog ist beiderseitig." Der Lotse soll in Form einer wachsamen Gemeinschaft also an Bord bleiben. Begünstigt wird diese Annahme auch durch die aufkommenden Anfasstechnologien, mit denen das iPhone derzeit so erfolgreich ist: "Allein zehn Prozent der menschlichen Gehirnleistung gehen auf den Pinzettengriff zurück." Martin Hirsch wundert sich, dass diese Erkenntnis von der IT-Branche erst jetzt berücksichtigt wird. Auf berührungssensiblen Bildschirmen sollen die Nutzer mit wenigen Handgriffen bald spielerisch Inhalte schaffen und dem System zur Verfügung stellen. So wird aus der Suchmaschine ein "Sparringspartner" und aus einem Informationsknopf ein "Knowledge Nugget". Wie auch immer man die Erkenntniszutaten des Internetgroßmarkts serviert: Wissen als Zeitwort ist ein länglicher Prozess. Im Moment sei die Maschine noch auf dem Stand eines Zweijährigen, sagen ihre Schöpfer. Sozialisiert werden soll sie demnächst im Internet, ihre Erziehung erfolgt dann durch die Nutzer. Als er Martin Hirsch mit seiner Scheibe zum ersten Mal gesehen habe, dachte Ralf von Grafenstein: "Das ist überfällig! Das wird kommen! Das muss raus!" Jetzt ist es da, klein, unschuldig und unscheinbar. Man findet es bei Google."
  7. Palm, F.: QVIZ : Query and context based visualization of time-spatial cultural dynamics (2007) 0.01
    0.010779224 = product of:
      0.043116897 = sum of:
        0.043116897 = sum of:
          0.008413256 = weight(_text_:a in 1289) [ClassicSimilarity], result of:
            0.008413256 = score(doc=1289,freq=10.0), product of:
              0.049223874 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.04269026 = queryNorm
              0.1709182 = fieldWeight in 1289, product of:
                3.1622777 = tf(freq=10.0), with freq of:
                  10.0 = termFreq=10.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=1289)
          0.034703642 = weight(_text_:22 in 1289) [ClassicSimilarity], result of:
            0.034703642 = score(doc=1289,freq=2.0), product of:
              0.149494 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04269026 = queryNorm
              0.23214069 = fieldWeight in 1289, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=1289)
      0.25 = coord(1/4)
    
    Abstract
    QVIZ will research and create a framework for visualizing and querying archival resources by a time-space interface based on maps and emergent knowledge structures. The framework will also integrate social software, such as wikis, in order to utilize knowledge in existing and new communities of practice. QVIZ will lead to improved information sharing and knowledge creation, easier access to information in a user-adapted context and innovative ways of exploring and visualizing materials over time, between countries and other administrative units. The common European framework for sharing and accessing archival information provided by the QVIZ project will open a considerably larger commercial market based on archival materials as well as a richer understanding of European history.
    Content
    Vortrag anlässlich des Workshops: "Extending the multilingual capacity of The European Library in the EDL project Stockholm, Swedish National Library, 22-23 November 2007".
  8. Osinska, V.; Bala, P.: New methods for visualization and improvement of classification schemes : the case of computer science (2010) 0.01
    0.010006163 = product of:
      0.040024653 = sum of:
        0.040024653 = sum of:
          0.0053210096 = weight(_text_:a in 3693) [ClassicSimilarity], result of:
            0.0053210096 = score(doc=3693,freq=4.0), product of:
              0.049223874 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.04269026 = queryNorm
              0.10809815 = fieldWeight in 3693, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.046875 = fieldNorm(doc=3693)
          0.034703642 = weight(_text_:22 in 3693) [ClassicSimilarity], result of:
            0.034703642 = score(doc=3693,freq=2.0), product of:
              0.149494 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04269026 = queryNorm
              0.23214069 = fieldWeight in 3693, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=3693)
      0.25 = coord(1/4)
    
    Abstract
    Generally, Computer Science (CS) classifications are inconsistent in taxonomy strategies. t is necessary to develop CS taxonomy research to combine its historical perspective, its current knowledge and its predicted future trends - including all breakthroughs in information and communication technology. In this paper we have analyzed the ACM Computing Classification System (CCS) by means of visualization maps. The important achievement of current work is an effective visualization of classified documents from the ACM Digital Library. From the technical point of view, the innovation lies in the parallel use of analysis units: (sub)classes and keywords as well as a spherical 3D information surface. We have compared both the thematic and semantic maps of classified documents and results presented in Table 1. Furthermore, the proposed new method is used for content-related evaluation of the original scheme. Summing up: we improved an original ACM classification in the Computer Science domain by means of visualization.
    Date
    22. 7.2010 19:36:46
    Type
    a
  9. Chen, C.: CiteSpace II : detecting and visualizing emerging trends and transient patterns in scientific literature (2006) 0.01
    0.009829691 = product of:
      0.039318763 = sum of:
        0.039318763 = sum of:
          0.010399062 = weight(_text_:a in 5272) [ClassicSimilarity], result of:
            0.010399062 = score(doc=5272,freq=22.0), product of:
              0.049223874 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.04269026 = queryNorm
              0.21126054 = fieldWeight in 5272, product of:
                4.690416 = tf(freq=22.0), with freq of:
                  22.0 = termFreq=22.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5272)
          0.028919702 = weight(_text_:22 in 5272) [ClassicSimilarity], result of:
            0.028919702 = score(doc=5272,freq=2.0), product of:
              0.149494 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04269026 = queryNorm
              0.19345059 = fieldWeight in 5272, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5272)
      0.25 = coord(1/4)
    
    Abstract
    This article describes the latest development of a generic approach to detecting and visualizing emerging trends and transient patterns in scientific literature. The work makes substantial theoretical and methodological contributions to progressive knowledge domain visualization. A specialty is conceptualized and visualized as a time-variant duality between two fundamental concepts in information science: research fronts and intellectual bases. A research front is defined as an emergent and transient grouping of concepts and underlying research issues. The intellectual base of a research front is its citation and co-citation footprint in scientific literature - an evolving network of scientific publications cited by research-front concepts. Kleinberg's (2002) burst-detection algorithm is adapted to identify emergent research-front concepts. Freeman's (1979) betweenness centrality metric is used to highlight potential pivotal points of paradigm shift over time. Two complementary visualization views are designed and implemented: cluster views and time-zone views. The contributions of the approach are that (a) the nature of an intellectual base is algorithmically and temporally identified by emergent research-front terms, (b) the value of a co-citation cluster is explicitly interpreted in terms of research-front concepts, and (c) visually prominent and algorithmically detected pivotal points substantially reduce the complexity of a visualized network. The modeling and visualization process is implemented in CiteSpace II, a Java application, and applied to the analysis of two research fields: mass extinction (1981-2004) and terrorism (1990-2003). Prominent trends and pivotal points in visualized networks were verified in collaboration with domain experts, who are the authors of pivotal-point articles. Practical implications of the work are discussed. A number of challenges and opportunities for future studies are identified.
    Date
    22. 7.2006 16:11:05
    Type
    a
  10. Eibl, M.: Visualisierung im Document Retrieval : Theoretische und praktische Zusammenführung von Softwareergonomie und Graphik Design (2000) 0.01
    0.00950153 = product of:
      0.03800612 = sum of:
        0.03800612 = weight(_text_:da in 2802) [ClassicSimilarity], result of:
          0.03800612 = score(doc=2802,freq=2.0), product of:
            0.20483522 = queryWeight, product of:
              4.7981725 = idf(docFreq=990, maxDocs=44218)
              0.04269026 = queryNorm
            0.18554485 = fieldWeight in 2802, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.7981725 = idf(docFreq=990, maxDocs=44218)
              0.02734375 = fieldNorm(doc=2802)
      0.25 = coord(1/4)
    
    Content
    Ausgangspunkt der vorliegenden Arbeit ist die Fragestellung, wie Anwender von Document Retrieval-Systemen bei der Formulierung einer Anfrage unterstützt werden können. Bei der Beantwortung dieser Frage gilt es zunächst, die verschiedenen Modelle des Document Retrieval auf ihre Stärken und Schwächen hin zu untersuchen. Es werden hier drei exemplarische Retrieval-Modelle herausgegriffen, die sich in ihrer Recherchegrundlage und Relevanzbehandlung stark voneinander unterscheiden: das Boolesche, das probabilistische und das vage Retrieval-Modell. Um den Recherchezugang beim Einsatz dieser Retrieval-Modelle anwenderfreundlich zu gestalten, wird auf das Mittel der Visualisierung zurückgegriffen. In ihr wird eine gute Möglichkeit gesehen, mit komplexer Information umzugehen, da sie im Gegensatz zu textbasierten Präsentationsformen der optisch ausgerichteten menschlichen Kognition sehr weit entgegenkommen kann. Jedoch zeigen zahlreiche Beispiele, dass Visualisierung keinesfalls als Allheilmittel gelten kann, sondern im Gegenteil sogar zu einer Komplexitätssteigerung der Materie führen kann. Eine genauere Analyse bereits bestehender Ansätze weist denn auch softwareergonomische Probleme nach und läßt die Neukonzeption einer Visualisierung aus softwareergonomischer Sicht notwendig erscheinen. Neben softwareergonomischen Überlegungen fließen auch Aspekte des Graphik Design in die Konzeption der Visualisierung mit ein. Dies ist insofern ein Novum, als beide Schulen einander bislang weitgehend ignorierten. Mit Hilfe des Graphik Design kann der Visualisierung eine ansprechendere Gestaltung gegeben werden, die eine erhöhte Anwenderakzeptanz bewirkt. Um die Kooperation von Softwareergonomie und Graphik Design auf eine allgemeingültige Basis stellen zu können, wird auch eine theoretische Zusammenführung vorgeschlagen. Auf der Basis dieser Vorüberlegungen wird eine Visualisierung vorgestellt, welche Aspekte der drei Retrieval-Modelle integriert: In ihrer Grundkonzeption ermöglicht sie Boolesche Recherche. In zwei Erweiterungen werden ein probabilistisches Ranking-Verfahren sowie die Möglichkeit der Erweiterung der Ergebnismenge durch vage Methoden zur Verfügung gestellt. Optisch verfolgt die Viz, sualisierung ein minimalistisches Design. Es werden zu eingegebenen Suchkriterien sämtliche möglichen Kombinationen mit der entsprechenden Anzahl der gefundenen Dokumente angezeigt. Die Codierung der Kombinationen erfolgt dabei rein über die Farbgebung, die durch die Farben der enthaltenen Suchkriterien bestimmt wird. Je nachdem, welche Erweiterung verwendet wird, verändert sich die Gestaltung der Visualisierung: Wird das probabilistische Rankigverfahren eingesetzt, so wird die Position der Kombinationen auf dem Bildschirm neu berechnet. Wird vages Retrieval eingesetzt, so erhalten die erweiterten Mengen ein neues graphisches Element. Um die Güte der Visualisierung zu ermitteln, wird sie in einem Nutzertest evaluiert. Es werden klassische Retrieval-Maße wie Recall und Precision ermittelt sowie in einem Fragebogen die Anwenderakzeptanz eruiert. Die Ergebnisse des Tests untermauern die zwei grundlegenden Thesen dieser Arbeit: Visualisierung ist ein sinnvolles Mittel, die Schwierigkeiten der Interaktion mit Document Retrieval-Systemen zu minimieren. Und die Integration von Softwareergonomie und Graphik Design ist nicht nur sowohl theoretisch als auch praktisch möglich, sondern in der Tat auch vorteilhaft
  11. Wu, K.-C.; Hsieh, T.-Y.: Affective choosing of clustering and categorization representations in e-book interfaces (2016) 0.01
    0.009303821 = product of:
      0.037215285 = sum of:
        0.037215285 = sum of:
          0.008295582 = weight(_text_:a in 3070) [ClassicSimilarity], result of:
            0.008295582 = score(doc=3070,freq=14.0), product of:
              0.049223874 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.04269026 = queryNorm
              0.1685276 = fieldWeight in 3070, product of:
                3.7416575 = tf(freq=14.0), with freq of:
                  14.0 = termFreq=14.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0390625 = fieldNorm(doc=3070)
          0.028919702 = weight(_text_:22 in 3070) [ClassicSimilarity], result of:
            0.028919702 = score(doc=3070,freq=2.0), product of:
              0.149494 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04269026 = queryNorm
              0.19345059 = fieldWeight in 3070, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=3070)
      0.25 = coord(1/4)
    
    Abstract
    Purpose - The purpose of this paper is to investigate user experiences with a touch-wall interface featuring both clustering and categorization representations of available e-books in a public library to understand human information interactions under work-focused and recreational contexts. Design/methodology/approach - Researchers collected questionnaires from 251 New Taipei City Library visitors who used the touch-wall interface to search for new titles. The authors applied structural equation modelling to examine relationships among hedonic/utilitarian needs, clustering and categorization representations, perceived ease of use (EU) and the extent to which users experienced anxiety and uncertainty (AU) while interacting with the interface. Findings - Utilitarian users who have an explicit idea of what they intend to find tend to prefer the categorization interface. A hedonic-oriented user tends to prefer clustering interfaces. Users reported EU regardless of which interface they engaged with. Results revealed that use of the clustering interface had a negative correlation with AU. Users that seek to satisfy utilitarian needs tended to emphasize the importance of perceived EU, whilst pleasure-seeking users were a little more tolerant of anxiety or uncertainty. Originality/value - The Online Public Access Catalogue (OPAC) encourages library visitors to borrow digital books through the implementation of an information visualization system. This situation poses an opportunity to validate uses and gratification theory. People with hedonic/utilitarian needs displayed different risk-control attitudes and affected uncertainty using the interface. Knowledge about user interaction with such interfaces is vital when launching the development of a new OPAC.
    Date
    20. 1.2015 18:30:22
    Type
    a
  12. Osinska, V.; Kowalska, M.; Osinski, Z.: ¬The role of visualization in the shaping and exploration of the individual information space : part 1 (2018) 0.01
    0.008982687 = product of:
      0.03593075 = sum of:
        0.03593075 = sum of:
          0.007011046 = weight(_text_:a in 4641) [ClassicSimilarity], result of:
            0.007011046 = score(doc=4641,freq=10.0), product of:
              0.049223874 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.04269026 = queryNorm
              0.14243183 = fieldWeight in 4641, product of:
                3.1622777 = tf(freq=10.0), with freq of:
                  10.0 = termFreq=10.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0390625 = fieldNorm(doc=4641)
          0.028919702 = weight(_text_:22 in 4641) [ClassicSimilarity], result of:
            0.028919702 = score(doc=4641,freq=2.0), product of:
              0.149494 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04269026 = queryNorm
              0.19345059 = fieldWeight in 4641, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=4641)
      0.25 = coord(1/4)
    
    Abstract
    Studies on the state and structure of digital knowledge concerning science generally relate to macro and meso scales. Supported by visualizations, these studies can deliver knowledge about emerging scientific fields or collaboration between countries, scientific centers, or groups of researchers. Analyses of individual activities or single scientific career paths are rarely presented and discussed. The authors decided to fill this gap and developed a web application for visualizing the scientific output of particular researchers. This free software based on bibliographic data from local databases, provides six layouts for analysis. Researchers can see the dynamic characteristics of their own writing activity, the time and place of publication, and the thematic scope of research problems. They can also identify cooperation networks, and consequently, study the dependencies and regularities in their own scientific activity. The current article presents the results of a study of the application's usability and functionality as well as attempts to define different user groups. A survey about the interface was sent to select researchers employed at Nicolaus Copernicus University. The results were used to answer the question as to whether such a specialized visualization tool can significantly augment the individual information space of the contemporary researcher.
    Date
    21.12.2018 17:22:13
    Type
    a
  13. Thissen, F.: Screen-Design-Manual : Communicating Effectively Through Multimedia (2003) 0.01
    0.008587609 = product of:
      0.034350436 = sum of:
        0.034350436 = sum of:
          0.0054307333 = weight(_text_:a in 1397) [ClassicSimilarity], result of:
            0.0054307333 = score(doc=1397,freq=6.0), product of:
              0.049223874 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.04269026 = queryNorm
              0.11032722 = fieldWeight in 1397, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1397)
          0.028919702 = weight(_text_:22 in 1397) [ClassicSimilarity], result of:
            0.028919702 = score(doc=1397,freq=2.0), product of:
              0.149494 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04269026 = queryNorm
              0.19345059 = fieldWeight in 1397, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1397)
      0.25 = coord(1/4)
    
    Abstract
    The "Screen Design Manual" provides designers of interactive media with a practical working guide for preparing and presenting information that is suitable for both their target groups and the media they are using. It describes background information and relationships, clarifies them with the help of examples, and encourages further development of the language of digital media. In addition to the basics of the psychology of perception and learning, ergonomics, communication theory, imagery research, and aesthetics, the book also explores the design of navigation and orientation elements. Guidelines and checklists, along with the unique presentation of the book, support the application of information in practice.
    Classification
    ST 253 Informatik / Monographien / Software und -entwicklung / Web-Programmierwerkzeuge (A-Z)
    Date
    22. 3.2008 14:29:25
    RVK
    ST 253 Informatik / Monographien / Software und -entwicklung / Web-Programmierwerkzeuge (A-Z)
  14. Wu, I.-C.; Vakkari, P.: Effects of subject-oriented visualization tools on search by novices and intermediates (2018) 0.01
    0.008338469 = product of:
      0.033353876 = sum of:
        0.033353876 = sum of:
          0.004434175 = weight(_text_:a in 4573) [ClassicSimilarity], result of:
            0.004434175 = score(doc=4573,freq=4.0), product of:
              0.049223874 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.04269026 = queryNorm
              0.090081796 = fieldWeight in 4573, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0390625 = fieldNorm(doc=4573)
          0.028919702 = weight(_text_:22 in 4573) [ClassicSimilarity], result of:
            0.028919702 = score(doc=4573,freq=2.0), product of:
              0.149494 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04269026 = queryNorm
              0.19345059 = fieldWeight in 4573, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=4573)
      0.25 = coord(1/4)
    
    Abstract
    This study explores how user subject knowledge influences search task processes and outcomes, as well as how search behavior is influenced by subject-oriented information visualization (IV) tools. To enable integrated searches, the proposed WikiMap + integrates search functions and IV tools (i.e., a topic network and hierarchical topic tree) and gathers information from Wikipedia pages and Google Search results. To evaluate the effectiveness of the proposed interfaces, we design subject-oriented tasks and adopt extended evaluation measures. We recruited 48 novices and 48 knowledgeable users, that is, intermediates, for the evaluation. Our results show that novices using the proposed interface demonstrate better search performance than intermediates using Wikipedia. We therefore conclude that our tools help close the gap between novices and intermediates in information searches. The results also show that intermediates can take advantage of the search tool by leveraging the IV tools to browse subtopics, and formulate better queries with less effort. We conclude that embedding the IV and the search tools in the interface can result in different search behavior but improved task performance. We provide implications to design search systems to include IV features adapted to user levels of subject knowledge to help them achieve better task performance.
    Date
    9.12.2018 16:22:25
    Type
    a
  15. Batorowska, H.; Kaminska-Czubala, B.: Information retrieval support : visualisation of the information space of a document (2014) 0.01
    0.007863753 = product of:
      0.03145501 = sum of:
        0.03145501 = sum of:
          0.008319249 = weight(_text_:a in 1444) [ClassicSimilarity], result of:
            0.008319249 = score(doc=1444,freq=22.0), product of:
              0.049223874 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.04269026 = queryNorm
              0.16900843 = fieldWeight in 1444, product of:
                4.690416 = tf(freq=22.0), with freq of:
                  22.0 = termFreq=22.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.03125 = fieldNorm(doc=1444)
          0.023135763 = weight(_text_:22 in 1444) [ClassicSimilarity], result of:
            0.023135763 = score(doc=1444,freq=2.0), product of:
              0.149494 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04269026 = queryNorm
              0.15476047 = fieldWeight in 1444, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=1444)
      0.25 = coord(1/4)
    
    Abstract
    Acquiring knowledge in any field involves information retrieval, i.e. searching the available documents to identify answers to the queries concerning the selected objects. Knowing the keywords which are names of the objects will enable situating the user's query in the information space organized as a thesaurus or faceted classification. Objectives: Identification the areas in the information space which correspond to gaps in the user's personal knowledge or in the domain knowledge might become useful in theory or practice. The aim of this paper is to present a realistic information-space model of a self-authored full-text document on information culture, indexed by the author of this article. Methodology: Having established the relations between the terms, particular modules (sets of terms connected by relations used in facet classification) are situated on a plain, similarly to a communication map. Conclusions drawn from the "journey" on the map, which is a visualization of the knowledge contained in the analysed document, are the crucial part of this paper. Results: The direct result of the research is the created model of information space visualization of a given document (book, article, website). The proposed procedure can practically be used as a new form of representation in order to map the contents of academic books and articles, beside the traditional index form, especially as an e-book auxiliary tool. In teaching, visualization of the information space of a document can be used to help students understand the issues of: classification, categorization and representation of new knowledge emerging in human mind.
    Source
    Knowledge organization in the 21st century: between historical patterns and future prospects. Proceedings of the Thirteenth International ISKO Conference 19-22 May 2014, Kraków, Poland. Ed.: Wieslaw Babik
    Type
    a
  16. Haller, S.H.M.: Mappingverfahren zur Wissensorganisation (2002) 0.01
    0.0072299256 = product of:
      0.028919702 = sum of:
        0.028919702 = product of:
          0.057839405 = sum of:
            0.057839405 = weight(_text_:22 in 3406) [ClassicSimilarity], result of:
              0.057839405 = score(doc=3406,freq=2.0), product of:
                0.149494 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04269026 = queryNorm
                0.38690117 = fieldWeight in 3406, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3406)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    30. 5.2010 16:22:35
  17. Spero, S.: LCSH is to thesaurus as doorbell is to mammal : visualizing structural problems in the Library of Congress Subject Headings (2008) 0.01
    0.0070381146 = product of:
      0.028152458 = sum of:
        0.028152458 = sum of:
          0.0050166957 = weight(_text_:a in 2659) [ClassicSimilarity], result of:
            0.0050166957 = score(doc=2659,freq=8.0), product of:
              0.049223874 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.04269026 = queryNorm
              0.10191591 = fieldWeight in 2659, product of:
                2.828427 = tf(freq=8.0), with freq of:
                  8.0 = termFreq=8.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.03125 = fieldNorm(doc=2659)
          0.023135763 = weight(_text_:22 in 2659) [ClassicSimilarity], result of:
            0.023135763 = score(doc=2659,freq=2.0), product of:
              0.149494 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04269026 = queryNorm
              0.15476047 = fieldWeight in 2659, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=2659)
      0.25 = coord(1/4)
    
    Abstract
    The Library of Congress Subject Headings (LCSH) has been developed over the course of more than a century, predating the semantic web by some time. Until the 1986, the only concept-toconcept relationship available was an undifferentiated "See Also" reference, which was used for both associative (RT) and hierarchical (BT/NT) connections. In that year, in preparation for the first release of the headings in machine readable MARC Authorities form, an attempt was made to automatically convert these "See Also" links into the standardized thesaural relations. Unfortunately, the rule used to determine the type of reference to generate relied on the presence of symmetric links to detect associatively related terms; "See Also" references that were only present in one of the related terms were assumed to be hierarchical. This left the process vulnerable to inconsistent use of references in the pre-conversion data, with a marked bias towards promoting relationships to hierarchical status. The Library of Congress was aware that the results of the conversion contained many inconsistencies, and intended to validate and correct the results over the course of time. Unfortunately, twenty years later, less than 40% of the converted records have been evaluated. The converted records, being the earliest encountered during the Library's cataloging activities, represent the most basic concepts within LCSH; errors in the syndetic structure for these records affect far more subordinate concepts than those nearer the periphery. Worse, a policy of patterning new headings after pre-existing ones leads to structural errors arising from the conversion process being replicated in these newer headings, perpetuating and exacerbating the errors. As the LCSH prepares for its second great conversion, from MARC to SKOS, it is critical to address these structural problems. As part of the work on converting the headings into SKOS, I have experimented with different visualizations of the tangled web of broader terms embedded in LCSH. This poster illustrates several of these renderings, shows how they can help users to judge which relationships might not be correct, and shows just exactly how Doorbells and Mammals are related.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
    Type
    a
  18. Graphic details : a scientific study of the importance of diagrams to science (2016) 0.01
    0.0068706805 = product of:
      0.027482722 = sum of:
        0.027482722 = sum of:
          0.010130901 = weight(_text_:a in 3035) [ClassicSimilarity], result of:
            0.010130901 = score(doc=3035,freq=58.0), product of:
              0.049223874 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.04269026 = queryNorm
              0.20581275 = fieldWeight in 3035, product of:
                7.615773 = tf(freq=58.0), with freq of:
                  58.0 = termFreq=58.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.0234375 = fieldNorm(doc=3035)
          0.017351821 = weight(_text_:22 in 3035) [ClassicSimilarity], result of:
            0.017351821 = score(doc=3035,freq=2.0), product of:
              0.149494 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04269026 = queryNorm
              0.116070345 = fieldWeight in 3035, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0234375 = fieldNorm(doc=3035)
      0.25 = coord(1/4)
    
    Abstract
    A PICTURE is said to be worth a thousand words. That metaphor might be expected to pertain a fortiori in the case of scientific papers, where a figure can brilliantly illuminate an idea that might otherwise be baffling. Papers with figures in them should thus be easier to grasp than those without. They should therefore reach larger audiences and, in turn, be more influential simply by virtue of being more widely read. But are they?
    Content
    Bill Howe and his colleagues at the University of Washington, in Seattle, decided to find out. First, they trained a computer algorithm to distinguish between various sorts of figures-which they defined as diagrams, equations, photographs, plots (such as bar charts and scatter graphs) and tables. They exposed their algorithm to between 400 and 600 images of each of these types of figure until it could distinguish them with an accuracy greater than 90%. Then they set it loose on the more-than-650,000 papers (containing more than 10m figures) stored on PubMed Central, an online archive of biomedical-research articles. To measure each paper's influence, they calculated its article-level Eigenfactor score-a modified version of the PageRank algorithm Google uses to provide the most relevant results for internet searches. Eigenfactor scoring gives a better measure than simply noting the number of times a paper is cited elsewhere, because it weights citations by their influence. A citation in a paper that is itself highly cited is worth more than one in a paper that is not.
    As the team describe in a paper posted (http://arxiv.org/abs/1605.04951) on arXiv, they found that figures did indeed matter-but not all in the same way. An average paper in PubMed Central has about one diagram for every three pages and gets 1.67 citations. Papers with more diagrams per page and, to a lesser extent, plots per page tended to be more influential (on average, a paper accrued two more citations for every extra diagram per page, and one more for every extra plot per page). By contrast, including photographs and equations seemed to decrease the chances of a paper being cited by others. That agrees with a study from 2012, whose authors counted (by hand) the number of mathematical expressions in over 600 biology papers and found that each additional equation per page reduced the number of citations a paper received by 22%. This does not mean that researchers should rush to include more diagrams in their next paper. Dr Howe has not shown what is behind the effect, which may merely be one of correlation, rather than causation. It could, for example, be that papers with lots of diagrams tend to be those that illustrate new concepts, and thus start a whole new field of inquiry. Such papers will certainly be cited a lot. On the other hand, the presence of equations really might reduce citations. Biologists (as are most of those who write and read the papers in PubMed Central) are notoriously mathsaverse. If that is the case, looking in a physics archive would probably produce a different result.
    Dr Howe and his colleagues do, however, believe that the study of diagrams can result in new insights. A figure showing new metabolic pathways in a cell, for example, may summarise hundreds of experiments. Since illustrations can convey important scientific concepts in this way, they think that browsing through related figures from different papers may help researchers come up with new theories. As Dr Howe puts it, "the unit of scientific currency is closer to the figure than to the paper." With this thought in mind, the team have created a website (viziometrics.org (http://viziometrics.org/) ) where the millions of images sorted by their program can be searched using key words. Their next plan is to extract the information from particular types of scientific figure, to create comprehensive "super" figures: a giant network of all the known chemical processes in a cell for example, or the best-available tree of life. At just one such superfigure per paper, though, the citation records of articles containing such all-embracing diagrams may very well undermine the correlation that prompted their creation in the first place. Call it the ultimate marriage of chart and science.
    Language
    a
    Type
    a
  19. Trunk, D.: Semantische Netze in Informationssystemen : Verbesserung der Suche durch Interaktion und Visualisierung (2005) 0.01
    0.005060948 = product of:
      0.020243792 = sum of:
        0.020243792 = product of:
          0.040487584 = sum of:
            0.040487584 = weight(_text_:22 in 2500) [ClassicSimilarity], result of:
              0.040487584 = score(doc=2500,freq=2.0), product of:
                0.149494 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04269026 = queryNorm
                0.2708308 = fieldWeight in 2500, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2500)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    30. 1.2007 18:22:41
  20. Information visualization in data mining and knowledge discovery (2002) 0.00
    0.004490735 = product of:
      0.01796294 = sum of:
        0.01796294 = sum of:
          0.0063950582 = weight(_text_:a in 1789) [ClassicSimilarity], result of:
            0.0063950582 = score(doc=1789,freq=52.0), product of:
              0.049223874 = queryWeight, product of:
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.04269026 = queryNorm
              0.12991782 = fieldWeight in 1789, product of:
                7.2111025 = tf(freq=52.0), with freq of:
                  52.0 = termFreq=52.0
                1.153047 = idf(docFreq=37942, maxDocs=44218)
                0.015625 = fieldNorm(doc=1789)
          0.011567881 = weight(_text_:22 in 1789) [ClassicSimilarity], result of:
            0.011567881 = score(doc=1789,freq=2.0), product of:
              0.149494 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04269026 = queryNorm
              0.07738023 = fieldWeight in 1789, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.015625 = fieldNorm(doc=1789)
      0.25 = coord(1/4)
    
    Date
    23. 3.2008 19:10:22
    Footnote
    Rez. in: JASIST 54(2003) no.9, S.905-906 (C.A. Badurek): "Visual approaches for knowledge discovery in very large databases are a prime research need for information scientists focused an extracting meaningful information from the ever growing stores of data from a variety of domains, including business, the geosciences, and satellite and medical imagery. This work presents a summary of research efforts in the fields of data mining, knowledge discovery, and data visualization with the goal of aiding the integration of research approaches and techniques from these major fields. The editors, leading computer scientists from academia and industry, present a collection of 32 papers from contributors who are incorporating visualization and data mining techniques through academic research as well application development in industry and government agencies. Information Visualization focuses upon techniques to enhance the natural abilities of humans to visually understand data, in particular, large-scale data sets. It is primarily concerned with developing interactive graphical representations to enable users to more intuitively make sense of multidimensional data as part of the data exploration process. It includes research from computer science, psychology, human-computer interaction, statistics, and information science. Knowledge Discovery in Databases (KDD) most often refers to the process of mining databases for previously unknown patterns and trends in data. Data mining refers to the particular computational methods or algorithms used in this process. The data mining research field is most related to computational advances in database theory, artificial intelligence and machine learning. This work compiles research summaries from these main research areas in order to provide "a reference work containing the collection of thoughts and ideas of noted researchers from the fields of data mining and data visualization" (p. 8). It addresses these areas in three main sections: the first an data visualization, the second an KDD and model visualization, and the last an using visualization in the knowledge discovery process. The seven chapters of Part One focus upon methodologies and successful techniques from the field of Data Visualization. Hoffman and Grinstein (Chapter 2) give a particularly good overview of the field of data visualization and its potential application to data mining. An introduction to the terminology of data visualization, relation to perceptual and cognitive science, and discussion of the major visualization display techniques are presented. Discussion and illustration explain the usefulness and proper context of such data visualization techniques as scatter plots, 2D and 3D isosurfaces, glyphs, parallel coordinates, and radial coordinate visualizations. Remaining chapters present the need for standardization of visualization methods, discussion of user requirements in the development of tools, and examples of using information visualization in addressing research problems.
    In 13 chapters, Part Two provides an introduction to KDD, an overview of data mining techniques, and examples of the usefulness of data model visualizations. The importance of visualization throughout the KDD process is stressed in many of the chapters. In particular, the need for measures of visualization effectiveness, benchmarking for identifying best practices, and the use of standardized sample data sets is convincingly presented. Many of the important data mining approaches are discussed in this complementary context. Cluster and outlier detection, classification techniques, and rule discovery algorithms are presented as the basic techniques common to the KDD process. The potential effectiveness of using visualization in the data modeling process are illustrated in chapters focused an using visualization for helping users understand the KDD process, ask questions and form hypotheses about their data, and evaluate the accuracy and veracity of their results. The 11 chapters of Part Three provide an overview of the KDD process and successful approaches to integrating KDD, data mining, and visualization in complementary domains. Rhodes (Chapter 21) begins this section with an excellent overview of the relation between the KDD process and data mining techniques. He states that the "primary goals of data mining are to describe the existing data and to predict the behavior or characteristics of future data of the same type" (p. 281). These goals are met by data mining tasks such as classification, regression, clustering, summarization, dependency modeling, and change or deviation detection. Subsequent chapters demonstrate how visualization can aid users in the interactive process of knowledge discovery by graphically representing the results from these iterative tasks. Finally, examples of the usefulness of integrating visualization and data mining tools in the domain of business, imagery and text mining, and massive data sets are provided. This text concludes with a thorough and useful 17-page index and lengthy yet integrating 17-page summary of the academic and industrial backgrounds of the contributing authors. A 16-page set of color inserts provide a better representation of the visualizations discussed, and a URL provided suggests that readers may view all the book's figures in color on-line, although as of this submission date it only provides access to a summary of the book and its contents. The overall contribution of this work is its focus an bridging two distinct areas of research, making it a valuable addition to the Morgan Kaufmann Series in Database Management Systems. The editors of this text have met their main goal of providing the first textbook integrating knowledge discovery, data mining, and visualization. Although it contributes greatly to our under- standing of the development and current state of the field, a major weakness of this text is that there is no concluding chapter to discuss the contributions of the sum of these contributed papers or give direction to possible future areas of research. "Integration of expertise between two different disciplines is a difficult process of communication and reeducation. Integrating data mining and visualization is particularly complex because each of these fields in itself must draw an a wide range of research experience" (p. 300). Although this work contributes to the crossdisciplinary communication needed to advance visualization in KDD, a more formal call for an interdisciplinary research agenda in a concluding chapter would have provided a more satisfying conclusion to a very good introductory text.
    With contributors almost exclusively from the computer science field, the intended audience of this work is heavily slanted towards a computer science perspective. However, it is highly readable and provides introductory material that would be useful to information scientists from a variety of domains. Yet, much interesting work in information visualization from other fields could have been included giving the work more of an interdisciplinary perspective to complement their goals of integrating work in this area. Unfortunately, many of the application chapters are these, shallow, and lack complementary illustrations of visualization techniques or user interfaces used. However, they do provide insight into the many applications being developed in this rapidly expanding field. The authors have successfully put together a highly useful reference text for the data mining and information visualization communities. Those interested in a good introduction and overview of complementary research areas in these fields will be satisfied with this collection of papers. The focus upon integrating data visualization with data mining complements texts in each of these fields, such as Advances in Knowledge Discovery and Data Mining (Fayyad et al., MIT Press) and Readings in Information Visualization: Using Vision to Think (Card et. al., Morgan Kauffman). This unique work is a good starting point for future interaction between researchers in the fields of data visualization and data mining and makes a good accompaniment for a course focused an integrating these areas or to the main reference texts in these fields."

Years

Languages

  • e 145
  • d 34
  • a 1
  • More… Less…

Types

  • a 157
  • el 32
  • m 11
  • x 4
  • s 2
  • b 1
  • p 1
  • r 1
  • More… Less…

Subjects

Classifications