Search (35 results, page 1 of 2)

  • × theme_ss:"Visualisierung"
  1. Thissen, F.: Screen-Design-Manual : Communicating Effectively Through Multimedia (2003) 0.02
    0.021699077 = product of:
      0.086796306 = sum of:
        0.086796306 = sum of:
          0.05596308 = weight(_text_:software in 1397) [ClassicSimilarity], result of:
            0.05596308 = score(doc=1397,freq=4.0), product of:
              0.18056466 = queryWeight, product of:
                3.9671519 = idf(docFreq=2274, maxDocs=44218)
                0.045514934 = queryNorm
              0.30993375 = fieldWeight in 1397, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.9671519 = idf(docFreq=2274, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1397)
          0.030833228 = weight(_text_:22 in 1397) [ClassicSimilarity], result of:
            0.030833228 = score(doc=1397,freq=2.0), product of:
              0.15938555 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.045514934 = queryNorm
              0.19345059 = fieldWeight in 1397, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1397)
      0.25 = coord(1/4)
    
    Classification
    ST 253 Informatik / Monographien / Software und -entwicklung / Web-Programmierwerkzeuge (A-Z)
    Date
    22. 3.2008 14:29:25
    RVK
    ST 253 Informatik / Monographien / Software und -entwicklung / Web-Programmierwerkzeuge (A-Z)
  2. Palm, F.: QVIZ : Query and context based visualization of time-spatial cultural dynamics (2007) 0.02
    0.021121528 = product of:
      0.08448611 = sum of:
        0.08448611 = sum of:
          0.047486246 = weight(_text_:software in 1289) [ClassicSimilarity], result of:
            0.047486246 = score(doc=1289,freq=2.0), product of:
              0.18056466 = queryWeight, product of:
                3.9671519 = idf(docFreq=2274, maxDocs=44218)
                0.045514934 = queryNorm
              0.2629875 = fieldWeight in 1289, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.9671519 = idf(docFreq=2274, maxDocs=44218)
                0.046875 = fieldNorm(doc=1289)
          0.03699987 = weight(_text_:22 in 1289) [ClassicSimilarity], result of:
            0.03699987 = score(doc=1289,freq=2.0), product of:
              0.15938555 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.045514934 = queryNorm
              0.23214069 = fieldWeight in 1289, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=1289)
      0.25 = coord(1/4)
    
    Abstract
    QVIZ will research and create a framework for visualizing and querying archival resources by a time-space interface based on maps and emergent knowledge structures. The framework will also integrate social software, such as wikis, in order to utilize knowledge in existing and new communities of practice. QVIZ will lead to improved information sharing and knowledge creation, easier access to information in a user-adapted context and innovative ways of exploring and visualizing materials over time, between countries and other administrative units. The common European framework for sharing and accessing archival information provided by the QVIZ project will open a considerably larger commercial market based on archival materials as well as a richer understanding of European history.
    Content
    Vortrag anlässlich des Workshops: "Extending the multilingual capacity of The European Library in the EDL project Stockholm, Swedish National Library, 22-23 November 2007".
  3. Osinska, V.; Kowalska, M.; Osinski, Z.: ¬The role of visualization in the shaping and exploration of the individual information space : part 1 (2018) 0.02
    0.017601274 = product of:
      0.070405096 = sum of:
        0.070405096 = sum of:
          0.03957187 = weight(_text_:software in 4641) [ClassicSimilarity], result of:
            0.03957187 = score(doc=4641,freq=2.0), product of:
              0.18056466 = queryWeight, product of:
                3.9671519 = idf(docFreq=2274, maxDocs=44218)
                0.045514934 = queryNorm
              0.21915624 = fieldWeight in 4641, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.9671519 = idf(docFreq=2274, maxDocs=44218)
                0.0390625 = fieldNorm(doc=4641)
          0.030833228 = weight(_text_:22 in 4641) [ClassicSimilarity], result of:
            0.030833228 = score(doc=4641,freq=2.0), product of:
              0.15938555 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.045514934 = queryNorm
              0.19345059 = fieldWeight in 4641, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=4641)
      0.25 = coord(1/4)
    
    Abstract
    Studies on the state and structure of digital knowledge concerning science generally relate to macro and meso scales. Supported by visualizations, these studies can deliver knowledge about emerging scientific fields or collaboration between countries, scientific centers, or groups of researchers. Analyses of individual activities or single scientific career paths are rarely presented and discussed. The authors decided to fill this gap and developed a web application for visualizing the scientific output of particular researchers. This free software based on bibliographic data from local databases, provides six layouts for analysis. Researchers can see the dynamic characteristics of their own writing activity, the time and place of publication, and the thematic scope of research problems. They can also identify cooperation networks, and consequently, study the dependencies and regularities in their own scientific activity. The current article presents the results of a study of the application's usability and functionality as well as attempts to define different user groups. A survey about the interface was sent to select researchers employed at Nicolaus Copernicus University. The results were used to answer the question as to whether such a specialized visualization tool can significantly augment the individual information space of the contemporary researcher.
    Date
    21.12.2018 17:22:13
  4. Eito Brun, R.: Retrieval effectiveness in software repositories : from faceted classifications to software visualization techniques (2006) 0.02
    0.017135125 = product of:
      0.0685405 = sum of:
        0.0685405 = product of:
          0.137081 = sum of:
            0.137081 = weight(_text_:software in 2515) [ClassicSimilarity], result of:
              0.137081 = score(doc=2515,freq=24.0), product of:
                0.18056466 = queryWeight, product of:
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.045514934 = queryNorm
                0.75917953 = fieldWeight in 2515, product of:
                  4.8989797 = tf(freq=24.0), with freq of:
                    24.0 = termFreq=24.0
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2515)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    The internal organization of large software projects requires an extraordinary effort in the development and maintenance of repositories made up of software artifacts (business components, data models, functional and technical documentation, etc.). During the software development process, different artifacts are created to help users in the transfer of knowledge and enable communication between workers and teams. The storage, maintenance and publication of these artifacts in knowledge bases - usually referred to as "software repositories" are a useful tool for future software development projects, as they contain the collective, learned experience of the teams and provide the basis to estimate and reuse the work completed in the past. Different techniques similar to those used by the library community have been used in the past to organize these software repositories and help users in the difficult task or identifying and retrieving artifacts (software and documentation). These techniques include software classification - with a special emphasis on faceted classifications, keyword-based retrieval and formal method techniques. The paper discusses the different knowledge organization techniques applied in these repositories to identify and retrieve software artifacts and ensure the reusability of software components and documentation at the different phases of the development process across different projects. An enumeration of the main approaches documented in specialized bibliography is provided.
  5. Cobo, M.J.; López-Herrera, A.G.; Herrera-Viedma, E.; Herrera, F.: Science mapping software tools : review, analysis, and cooperative study among tools (2011) 0.01
    0.013708099 = product of:
      0.054832395 = sum of:
        0.054832395 = product of:
          0.10966479 = sum of:
            0.10966479 = weight(_text_:software in 4486) [ClassicSimilarity], result of:
              0.10966479 = score(doc=4486,freq=6.0), product of:
                0.18056466 = queryWeight, product of:
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.045514934 = queryNorm
                0.6073436 = fieldWeight in 4486, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4486)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    Science mapping aims to build bibliometric maps that describe how specific disciplines, scientific domains, or research fields are conceptually, intellectually, and socially structured. Different techniques and software tools have been proposed to carry out science mapping analysis. The aim of this article is to review, analyze, and compare some of these software tools, taking into account aspects such as the bibliometric techniques available and the different kinds of analysis.
  6. Pejtersen, A.M.: Implications of users' value perception for the design of a bibliographic retrieval system (1986) 0.01
    0.011871561 = product of:
      0.047486246 = sum of:
        0.047486246 = product of:
          0.09497249 = sum of:
            0.09497249 = weight(_text_:software in 2961) [ClassicSimilarity], result of:
              0.09497249 = score(doc=2961,freq=2.0), product of:
                0.18056466 = queryWeight, product of:
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.045514934 = queryNorm
                0.525975 = fieldWeight in 2961, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.09375 = fieldNorm(doc=2961)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Source
    Empirical foundation of information and software science. Ed.: J.C. Agarwal u. P. Zunde
  7. Eibl, M.: Recherche in elektronischen Bibliothekskatalogen (1999) 0.01
    0.0098929675 = product of:
      0.03957187 = sum of:
        0.03957187 = product of:
          0.07914374 = sum of:
            0.07914374 = weight(_text_:software in 3465) [ClassicSimilarity], result of:
              0.07914374 = score(doc=3465,freq=2.0), product of:
                0.18056466 = queryWeight, product of:
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.045514934 = queryNorm
                0.43831247 = fieldWeight in 3465, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3465)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Source
    Software
  8. Sieber, W.: Thesaurus-Arbeit versus Informationsvisualisierung : Analyse und Evaluation am Maßstab der Usability (2007) 0.01
    0.008394462 = product of:
      0.03357785 = sum of:
        0.03357785 = product of:
          0.0671557 = sum of:
            0.0671557 = weight(_text_:software in 1457) [ClassicSimilarity], result of:
              0.0671557 = score(doc=1457,freq=4.0), product of:
                0.18056466 = queryWeight, product of:
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.045514934 = queryNorm
                0.3719205 = fieldWeight in 1457, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1457)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    Traditionell werden Thesauren in Listen dargestellt, doch die Existenz von Software legt nahe, die Möglichkeiten moderner Informationsvisualisierung auszuschöpfen.Wolfram Sieber evaluiert, ob und wie Software zur Thesaurus-Visualisierung über die reine Listendarstellung hinausgeht. Besonderes Augenmerk richtet er darauf, wie gut dabei die Chancen der Visualisierung - Stichworte prä-/attentive Wahrnehmung und Change Blindness - genutzt werden, bzw. wie gegen sie verstoßen und der Nutzer dadurch in seiner Arbeit behindert wird.Zugunsten einer soliden Entscheidungsgrundlage hat Sieber eine detaillierte Liste von Prüfkriterien zusammengestellt und sieben verschiedene Thesaurus-Visualisierer daran gemessen. Diese Liste kann als Grundlage für weitergehende Evaluierungen verwendet werden, bietet sich aber vor allem als Richtschnur bei der Entwicklung eigener Thesaurus- und Graph-Visualisierer an. Anwendern und interessierten Laien hilft sie, zur Auswahl stehende Graph-Visualisierer aus fachlicher Sicht, wissenschaftlich abgesichert zu beurteilen. Dieses Werk richtet sich an Entscheider, Entwickler und Anwender aus Informationswissenschaft und -praxis, Informatik, Dokumentation.
  9. Yi, K.; Chan, L.M.: ¬A visualization software tool for Library of Congress Subject Headings (2008) 0.01
    0.008394462 = product of:
      0.03357785 = sum of:
        0.03357785 = product of:
          0.0671557 = sum of:
            0.0671557 = weight(_text_:software in 2503) [ClassicSimilarity], result of:
              0.0671557 = score(doc=2503,freq=4.0), product of:
                0.18056466 = queryWeight, product of:
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.045514934 = queryNorm
                0.3719205 = fieldWeight in 2503, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2503)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Content
    The aim of this study is to develop a software tool, VisuaLCSH, for effective searching, browsing, and maintenance of LCSH. This tool enables visualizing subject headings and hierarchical structures implied and embedded in LCSH. A conceptual framework for converting the hierarchical structure of headings in LCSH to an explicit tree structure is proposed, described, and implemented. The highlights of VisuaLCSH are summarized below: 1) revealing multiple aspects of a heading; 2) normalizing the hierarchical relationships in LCSH; 3) showing multi-level hierarchies in LCSH sub-trees; 4) improving the navigational function of LCSH in retrieval; and 5) enabling the implementation of generic search, i.e., the 'exploding' feature, in searching LCSH.
  10. Petersen, A.; Münch, V.: STN® AnaVist(TM) holt verborgenes Wissen aus Recherche-Ergebnissen : Neue Software analysiert und visualisiert Marktaufteilung, Forschung und Patentaktivitäten (2005) 0.01
    0.007914375 = product of:
      0.0316575 = sum of:
        0.0316575 = product of:
          0.063315 = sum of:
            0.063315 = weight(_text_:software in 3984) [ClassicSimilarity], result of:
              0.063315 = score(doc=3984,freq=8.0), product of:
                0.18056466 = queryWeight, product of:
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.045514934 = queryNorm
                0.35064998 = fieldWeight in 3984, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3984)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    "Im 21. Jahrhundert ist die entscheidende Herausforderung an Informationsdienstleister nicht, Informationen zugänglich, sondern sie optimal nutzbar zu machen", sagt Sabine Brünger-Weilandt, Geschäftsführerin von FIZ Karlsruhe, das den Online-Dienst STN International in internationaler Kooperation betreibt. Informationsprofis, so Brünger-Weilandt weiter, bräuchten hockentwickelte Software für strategisches Informationsmanagement. Als "Antwort auf diesen Bedarf" hat STN International eine neue Software zur Analyse und Visualisierung (A&V) von Rechercheergebnissen aus STN-Datenbanken entwickelt. STN® AnaVistT(TM) wurde auf der DGI Online-Tagung Ende Mai in Frankfurt am Main und auf Benutzertreffen in Frankfurt am Main, München und Essen vorgestellt. Seit 18. Juli 2005 ist das neue A&V-Werkzeug für die öffentliche Nutzung freigegeben (www.stn-international.de).
    Die wichtigsten Funktionen von STN AnaVist sind: - Inhalte aus mehreren Datenbanken sind gleichzeitig auswertbar - Nutzer können Daten aus unterschiedlichen Ouellen suchen, analysieren und visualisieren, u.a. aus der Chemiedatenbank CAplusSM, der Patentdatenbank PCTFULL, und US-amerikanischen Volltextdatenbanken. - Einzigartige Beziehungen zwischen Datenelementen-nur STN AnaVist bietet die Möglichkeit, Beziehungen zwischen sieben unterschiedlichen Feldern aus Datenbankdokumenten - z.B., Firmen, Erfindern, Veröffentlichungsjahren und Konzepten-darzustellen. - Gruppierung und Bereinigung von Daten - vor der Analyse werden Firmen und ihre unterschiedlichen Namensvarianten von einem "Company Name Thesaurus" zusammengefasst. - Konzept-Standardisierung - Durch das CAS-Vokabular werden Fachbegriffe datenbankübergreifend standardisiert, so dass weniger Streuung auftritt. - Interaktive Präsentation der Beziehungen zwischen Daten und Diagrammenwährend der Auswertung können Daten zum besseren Erkennen der Beziehungen farblich hervorgehoben werden. - Flexible Erstellung der auszuwertenden Rechercheergebnisse - Rechercheergebnisse, die als Ausgangsdatensatz für die Analyse verwendet werden sollen, können auf zwei Arten gewonnen werden: zum einen über die in STN® AnaVist(TM) integrierte Konzept-Suchfunktion, zum anderen durch problemlose Übernahme von Suchergebnissen aus der bewährten Software STN Express® with Discover! TM Analysis Edition, Version 8.0
  11. Haller, S.H.M.: Mappingverfahren zur Wissensorganisation (2002) 0.01
    0.007708307 = product of:
      0.030833228 = sum of:
        0.030833228 = product of:
          0.061666455 = sum of:
            0.061666455 = weight(_text_:22 in 3406) [ClassicSimilarity], result of:
              0.061666455 = score(doc=3406,freq=2.0), product of:
                0.15938555 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045514934 = queryNorm
                0.38690117 = fieldWeight in 3406, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3406)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    30. 5.2010 16:22:35
  12. Platis, N. et al.: Visualization of uncertainty in tag clouds (2016) 0.01
    0.007708307 = product of:
      0.030833228 = sum of:
        0.030833228 = product of:
          0.061666455 = sum of:
            0.061666455 = weight(_text_:22 in 2755) [ClassicSimilarity], result of:
              0.061666455 = score(doc=2755,freq=2.0), product of:
                0.15938555 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045514934 = queryNorm
                0.38690117 = fieldWeight in 2755, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2755)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    1. 2.2016 18:25:22
  13. Eckert, K.: Thesaurus analysis and visualization in semantic search applications (2007) 0.01
    0.006995385 = product of:
      0.02798154 = sum of:
        0.02798154 = product of:
          0.05596308 = sum of:
            0.05596308 = weight(_text_:software in 3222) [ClassicSimilarity], result of:
              0.05596308 = score(doc=3222,freq=4.0), product of:
                0.18056466 = queryWeight, product of:
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.045514934 = queryNorm
                0.30993375 = fieldWeight in 3222, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3222)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Content
    Vgl. unter: http://ki.informatik.uni-mannheim.de/fileadmin/publication/Eckert07Thesis.pdf. Für die Software vgl.: http://www.semtinel.org. Zur Beschreibung der Software: https://ub-madoc.bib.uni-mannheim.de/29611/.
  14. Sieber, W.: Visualisierung von Thesaurus-Strukturen unter besonderer Berücksichtigung eines Hyperbolic Tree Views (2004) 0.01
    0.006925077 = product of:
      0.027700309 = sum of:
        0.027700309 = product of:
          0.055400617 = sum of:
            0.055400617 = weight(_text_:software in 1456) [ClassicSimilarity], result of:
              0.055400617 = score(doc=1456,freq=2.0), product of:
                0.18056466 = queryWeight, product of:
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.045514934 = queryNorm
                0.30681872 = fieldWeight in 1456, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1456)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    Wenige Software-Werkzeuge zur Visualisierung von Thesaurus-Strukturen sind über das prototypische Stadium hinaus gelangt. Konkrete Implementationen entsprechender Anwendungen werden evaluiert. Aus den Gebieten Thesaurus-Entwicklung, Visualisierung, Interaktion, Informations- und Graphenvisualisierung werden Kriterien ermittelt, an denen die Implementierungen gemessen werden. Ein generisches Werkzeug zur Visualisierung dreidimensionaler hyperbolischer Bäume wird dazu verwendet, die Strukturen eines Thesaurus darzustellen. Dieses Werkzeug wird mit in die Evaluation einbezogen. Es hat sich gezeigt, dass die Visualisierung der Thesaurus-Strukturen auf Basis weniger unterschiedlicher Verfahren erfolgt. Der generische Visualisierer ist für Thesaurus-Strukturen prinzipiell geeignet, doch fehlen ihm spezifische Möglichkeiten der Domäne Thesaurus.
  15. Börner, K.: Atlas of knowledge : anyone can map (2015) 0.01
    0.0065407157 = product of:
      0.026162863 = sum of:
        0.026162863 = product of:
          0.052325726 = sum of:
            0.052325726 = weight(_text_:22 in 3355) [ClassicSimilarity], result of:
              0.052325726 = score(doc=3355,freq=4.0), product of:
                0.15938555 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045514934 = queryNorm
                0.32829654 = fieldWeight in 3355, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3355)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    22. 1.2017 16:54:03
    22. 1.2017 17:10:56
  16. Zhu, B.; Chen, H.: Information visualization (2004) 0.01
    0.005997293 = product of:
      0.023989173 = sum of:
        0.023989173 = product of:
          0.047978345 = sum of:
            0.047978345 = weight(_text_:software in 4276) [ClassicSimilarity], result of:
              0.047978345 = score(doc=4276,freq=6.0), product of:
                0.18056466 = queryWeight, product of:
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.045514934 = queryNorm
                0.26571283 = fieldWeight in 4276, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=4276)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    Advanced technology has resulted in the generation of about one million terabytes of information every year. Ninety-reine percent of this is available in digital format (Keim, 2001). More information will be generated in the next three years than was created during all of previous human history (Keim, 2001). Collecting information is no longer a problem, but extracting value from information collections has become progressively more difficult. Various search engines have been developed to make it easier to locate information of interest, but these work well only for a person who has a specific goal and who understands what and how information is stored. This usually is not the Gase. Visualization was commonly thought of in terms of representing human mental processes (MacEachren, 1991; Miller, 1984). The concept is now associated with the amplification of these mental processes (Card, Mackinlay, & Shneiderman, 1999). Human eyes can process visual cues rapidly, whereas advanced information analysis techniques transform the computer into a powerful means of managing digitized information. Visualization offers a link between these two potent systems, the human eye and the computer (Gershon, Eick, & Card, 1998), helping to identify patterns and to extract insights from large amounts of information. The identification of patterns is important because it may lead to a scientific discovery, an interpretation of clues to solve a crime, the prediction of catastrophic weather, a successful financial investment, or a better understanding of human behavior in a computermediated environment. Visualization technology shows considerable promise for increasing the value of large-scale collections of information, as evidenced by several commercial applications of TreeMap (e.g., http://www.smartmoney.com) and Hyperbolic tree (e.g., http://www.inxight.com) to visualize large-scale hierarchical structures. Although the proliferation of visualization technologies dates from the 1990s where sophisticated hardware and software made increasingly faster generation of graphical objects possible, the role of visual aids in facilitating the construction of mental images has a long history. Visualization has been used to communicate ideas, to monitor trends implicit in data, and to explore large volumes of data for hypothesis generation. Imagine traveling to a strange place without a map, having to memorize physical and chemical properties of an element without Mendeleyev's periodic table, trying to understand the stock market without statistical diagrams, or browsing a collection of documents without interactive visual aids. A collection of information can lose its value simply because of the effort required for exhaustive exploration. Such frustrations can be overcome by visualization.
    Visualization can be classified as scientific visualization, software visualization, or information visualization. Although the data differ, the underlying techniques have much in common. They use the same elements (visual cues) and follow the same rules of combining visual cues to deliver patterns. They all involve understanding human perception (Encarnacao, Foley, Bryson, & Feiner, 1994) and require domain knowledge (Tufte, 1990). Because most decisions are based an unstructured information, such as text documents, Web pages, or e-mail messages, this chapter focuses an the visualization of unstructured textual documents. The chapter reviews information visualization techniques developed over the last decade and examines how they have been applied in different domains. The first section provides the background by describing visualization history and giving overviews of scientific, software, and information visualization as well as the perceptual aspects of visualization. The next section assesses important visualization techniques that convert abstract information into visual objects and facilitate navigation through displays an a computer screen. It also explores information analysis algorithms that can be applied to identify or extract salient visualizable structures from collections of information. Information visualization systems that integrate different types of technologies to address problems in different domains are then surveyed; and we move an to a survey and critique of visualization system evaluation studies. The chapter concludes with a summary and identification of future research directions.
  17. Leydesdorff, L.; Persson, O.: Mapping the geography of science : distribution patterns and networks of relations among cities and institutes (2010) 0.01
    0.0059357807 = product of:
      0.023743123 = sum of:
        0.023743123 = product of:
          0.047486246 = sum of:
            0.047486246 = weight(_text_:software in 3704) [ClassicSimilarity], result of:
              0.047486246 = score(doc=3704,freq=2.0), product of:
                0.18056466 = queryWeight, product of:
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.045514934 = queryNorm
                0.2629875 = fieldWeight in 3704, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3704)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    Using Google Earth, Google Maps, and/or network visualization programs such as Pajek, one can overlay the network of relations among addresses in scientific publications onto the geographic map. The authors discuss the pros and cons of various options, and provide software (freeware) for bridging existing gaps between the Science Citation Indices (Thomson Reuters) and Scopus (Elsevier), on the one hand, and these various visualization tools on the other. At the level of city names, the global map can be drawn reliably on the basis of the available address information. At the level of the names of organizations and institutes, there are problems of unification both in the ISI databases and with Scopus. Pajek enables a combination of visualization and statistical analysis, whereas the Google Maps and its derivatives provide superior tools on the Internet.
  18. Aufreiter, M.: Informationsvisualisierung und Navigation im Semantic Web (2008) 0.01
    0.0059357807 = product of:
      0.023743123 = sum of:
        0.023743123 = product of:
          0.047486246 = sum of:
            0.047486246 = weight(_text_:software in 4711) [ClassicSimilarity], result of:
              0.047486246 = score(doc=4711,freq=2.0), product of:
                0.18056466 = queryWeight, product of:
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.045514934 = queryNorm
                0.2629875 = fieldWeight in 4711, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4711)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Content
    Bachelorarbeit, eingereicht am Fachhochschul-Bachelorstudiengang Software Engineering in Hagenberg
  19. Kraker, P.; Kittel, C,; Enkhbayar, A.: Open Knowledge Maps : creating a visual interface to the world's scientific knowledge based on natural language processing (2016) 0.01
    0.0059357807 = product of:
      0.023743123 = sum of:
        0.023743123 = product of:
          0.047486246 = sum of:
            0.047486246 = weight(_text_:software in 3205) [ClassicSimilarity], result of:
              0.047486246 = score(doc=3205,freq=2.0), product of:
                0.18056466 = queryWeight, product of:
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.045514934 = queryNorm
                0.2629875 = fieldWeight in 3205, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9671519 = idf(docFreq=2274, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3205)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Abstract
    The goal of Open Knowledge Maps is to create a visual interface to the world's scientific knowledge. The base for this visual interface consists of so-called knowledge maps, which enable the exploration of existing knowledge and the discovery of new knowledge. Our open source knowledge mapping software applies a mixture of summarization techniques and similarity measures on article metadata, which are iteratively chained together. After processing, the representation is saved in a database for use in a web visualization. In the future, we want to create a space for collective knowledge mapping that brings together individuals and communities involved in exploration and discovery. We want to enable people to guide each other in their discovery by collaboratively annotating and modifying the automatically created maps.
  20. Trunk, D.: Semantische Netze in Informationssystemen : Verbesserung der Suche durch Interaktion und Visualisierung (2005) 0.01
    0.005395815 = product of:
      0.02158326 = sum of:
        0.02158326 = product of:
          0.04316652 = sum of:
            0.04316652 = weight(_text_:22 in 2500) [ClassicSimilarity], result of:
              0.04316652 = score(doc=2500,freq=2.0), product of:
                0.15938555 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045514934 = queryNorm
                0.2708308 = fieldWeight in 2500, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2500)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    30. 1.2007 18:22:41

Years

Languages

  • e 23
  • d 11
  • a 1
  • More… Less…

Types

  • a 20
  • x 7
  • el 6
  • m 6
  • s 1
  • More… Less…