Search (77 results, page 1 of 4)

  • × theme_ss:"Visualisierung"
  1. Salaba, A.; Mercun, T.; Aalberg, T.: Complexity of work families and entity-based visualization displays (2018) 0.05
    0.054321747 = product of:
      0.13580437 = sum of:
        0.105629526 = weight(_text_:bibliographic in 5184) [ClassicSimilarity], result of:
          0.105629526 = score(doc=5184,freq=8.0), product of:
            0.17541347 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.04505818 = queryNorm
            0.6021745 = fieldWeight in 5184, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5184)
        0.03017484 = product of:
          0.06034968 = sum of:
            0.06034968 = weight(_text_:data in 5184) [ClassicSimilarity], result of:
              0.06034968 = score(doc=5184,freq=6.0), product of:
                0.14247625 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.04505818 = queryNorm
                0.42357713 = fieldWeight in 5184, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5184)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Conceptual modeling of bibliographic data, including the FR models and the consolidated IFLA LRM, has provided an opportunity to shift focus to entities and relationships and to support hierarchical work-based exploration of bibliographic information. This paper reports on a study examining the complexity of a work's bibliographic family data and user interactions with data visualizations, compared to traditional displays. Findings suggest that the FRBR-based visual bibliographic information system supports work families of different complexities more equally than a traditional system. Differences between the two systems also show that the FRBR-based system was more effective especially for related-works and author-related tasks.
  2. Information visualization in data mining and knowledge discovery (2002) 0.05
    0.04807465 = product of:
      0.12018663 = sum of:
        0.037583813 = weight(_text_:readable in 1789) [ClassicSimilarity], result of:
          0.037583813 = score(doc=1789,freq=2.0), product of:
            0.2768342 = queryWeight, product of:
              6.1439276 = idf(docFreq=257, maxDocs=44218)
              0.04505818 = queryNorm
            0.1357629 = fieldWeight in 1789, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.1439276 = idf(docFreq=257, maxDocs=44218)
              0.015625 = fieldNorm(doc=1789)
        0.082602814 = sum of:
          0.070393294 = weight(_text_:data in 1789) [ClassicSimilarity], result of:
            0.070393294 = score(doc=1789,freq=100.0), product of:
              0.14247625 = queryWeight, product of:
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.04505818 = queryNorm
              0.49407038 = fieldWeight in 1789, product of:
                10.0 = tf(freq=100.0), with freq of:
                  100.0 = termFreq=100.0
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.015625 = fieldNorm(doc=1789)
          0.012209523 = weight(_text_:22 in 1789) [ClassicSimilarity], result of:
            0.012209523 = score(doc=1789,freq=2.0), product of:
              0.15778607 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04505818 = queryNorm
              0.07738023 = fieldWeight in 1789, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.015625 = fieldNorm(doc=1789)
      0.4 = coord(2/5)
    
    Date
    23. 3.2008 19:10:22
    Footnote
    Rez. in: JASIST 54(2003) no.9, S.905-906 (C.A. Badurek): "Visual approaches for knowledge discovery in very large databases are a prime research need for information scientists focused an extracting meaningful information from the ever growing stores of data from a variety of domains, including business, the geosciences, and satellite and medical imagery. This work presents a summary of research efforts in the fields of data mining, knowledge discovery, and data visualization with the goal of aiding the integration of research approaches and techniques from these major fields. The editors, leading computer scientists from academia and industry, present a collection of 32 papers from contributors who are incorporating visualization and data mining techniques through academic research as well application development in industry and government agencies. Information Visualization focuses upon techniques to enhance the natural abilities of humans to visually understand data, in particular, large-scale data sets. It is primarily concerned with developing interactive graphical representations to enable users to more intuitively make sense of multidimensional data as part of the data exploration process. It includes research from computer science, psychology, human-computer interaction, statistics, and information science. Knowledge Discovery in Databases (KDD) most often refers to the process of mining databases for previously unknown patterns and trends in data. Data mining refers to the particular computational methods or algorithms used in this process. The data mining research field is most related to computational advances in database theory, artificial intelligence and machine learning. This work compiles research summaries from these main research areas in order to provide "a reference work containing the collection of thoughts and ideas of noted researchers from the fields of data mining and data visualization" (p. 8). It addresses these areas in three main sections: the first an data visualization, the second an KDD and model visualization, and the last an using visualization in the knowledge discovery process. The seven chapters of Part One focus upon methodologies and successful techniques from the field of Data Visualization. Hoffman and Grinstein (Chapter 2) give a particularly good overview of the field of data visualization and its potential application to data mining. An introduction to the terminology of data visualization, relation to perceptual and cognitive science, and discussion of the major visualization display techniques are presented. Discussion and illustration explain the usefulness and proper context of such data visualization techniques as scatter plots, 2D and 3D isosurfaces, glyphs, parallel coordinates, and radial coordinate visualizations. Remaining chapters present the need for standardization of visualization methods, discussion of user requirements in the development of tools, and examples of using information visualization in addressing research problems.
    In 13 chapters, Part Two provides an introduction to KDD, an overview of data mining techniques, and examples of the usefulness of data model visualizations. The importance of visualization throughout the KDD process is stressed in many of the chapters. In particular, the need for measures of visualization effectiveness, benchmarking for identifying best practices, and the use of standardized sample data sets is convincingly presented. Many of the important data mining approaches are discussed in this complementary context. Cluster and outlier detection, classification techniques, and rule discovery algorithms are presented as the basic techniques common to the KDD process. The potential effectiveness of using visualization in the data modeling process are illustrated in chapters focused an using visualization for helping users understand the KDD process, ask questions and form hypotheses about their data, and evaluate the accuracy and veracity of their results. The 11 chapters of Part Three provide an overview of the KDD process and successful approaches to integrating KDD, data mining, and visualization in complementary domains. Rhodes (Chapter 21) begins this section with an excellent overview of the relation between the KDD process and data mining techniques. He states that the "primary goals of data mining are to describe the existing data and to predict the behavior or characteristics of future data of the same type" (p. 281). These goals are met by data mining tasks such as classification, regression, clustering, summarization, dependency modeling, and change or deviation detection. Subsequent chapters demonstrate how visualization can aid users in the interactive process of knowledge discovery by graphically representing the results from these iterative tasks. Finally, examples of the usefulness of integrating visualization and data mining tools in the domain of business, imagery and text mining, and massive data sets are provided. This text concludes with a thorough and useful 17-page index and lengthy yet integrating 17-page summary of the academic and industrial backgrounds of the contributing authors. A 16-page set of color inserts provide a better representation of the visualizations discussed, and a URL provided suggests that readers may view all the book's figures in color on-line, although as of this submission date it only provides access to a summary of the book and its contents. The overall contribution of this work is its focus an bridging two distinct areas of research, making it a valuable addition to the Morgan Kaufmann Series in Database Management Systems. The editors of this text have met their main goal of providing the first textbook integrating knowledge discovery, data mining, and visualization. Although it contributes greatly to our under- standing of the development and current state of the field, a major weakness of this text is that there is no concluding chapter to discuss the contributions of the sum of these contributed papers or give direction to possible future areas of research. "Integration of expertise between two different disciplines is a difficult process of communication and reeducation. Integrating data mining and visualization is particularly complex because each of these fields in itself must draw an a wide range of research experience" (p. 300). Although this work contributes to the crossdisciplinary communication needed to advance visualization in KDD, a more formal call for an interdisciplinary research agenda in a concluding chapter would have provided a more satisfying conclusion to a very good introductory text.
    With contributors almost exclusively from the computer science field, the intended audience of this work is heavily slanted towards a computer science perspective. However, it is highly readable and provides introductory material that would be useful to information scientists from a variety of domains. Yet, much interesting work in information visualization from other fields could have been included giving the work more of an interdisciplinary perspective to complement their goals of integrating work in this area. Unfortunately, many of the application chapters are these, shallow, and lack complementary illustrations of visualization techniques or user interfaces used. However, they do provide insight into the many applications being developed in this rapidly expanding field. The authors have successfully put together a highly useful reference text for the data mining and information visualization communities. Those interested in a good introduction and overview of complementary research areas in these fields will be satisfied with this collection of papers. The focus upon integrating data visualization with data mining complements texts in each of these fields, such as Advances in Knowledge Discovery and Data Mining (Fayyad et al., MIT Press) and Readings in Information Visualization: Using Vision to Think (Card et. al., Morgan Kauffman). This unique work is a good starting point for future interaction between researchers in the fields of data visualization and data mining and makes a good accompaniment for a course focused an integrating these areas or to the main reference texts in these fields."
    LCSH
    Data mining
    RSWK
    Visualisierung / Computergraphik / Data Mining
    Data Mining / Visualisierung / Aufsatzsammlung (BVB)
    Series
    Morgan Kaufmann series in data management systems
    Subject
    Visualisierung / Computergraphik / Data Mining
    Data Mining / Visualisierung / Aufsatzsammlung (BVB)
    Data mining
    Theme
    Data Mining
  3. Spero, S.: LCSH is to thesaurus as doorbell is to mammal : visualizing structural problems in the Library of Congress Subject Headings (2008) 0.05
    0.04779876 = product of:
      0.1194969 = sum of:
        0.075167626 = weight(_text_:readable in 2659) [ClassicSimilarity], result of:
          0.075167626 = score(doc=2659,freq=2.0), product of:
            0.2768342 = queryWeight, product of:
              6.1439276 = idf(docFreq=257, maxDocs=44218)
              0.04505818 = queryNorm
            0.2715258 = fieldWeight in 2659, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.1439276 = idf(docFreq=257, maxDocs=44218)
              0.03125 = fieldNorm(doc=2659)
        0.044329274 = sum of:
          0.01991023 = weight(_text_:data in 2659) [ClassicSimilarity], result of:
            0.01991023 = score(doc=2659,freq=2.0), product of:
              0.14247625 = queryWeight, product of:
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.04505818 = queryNorm
              0.1397442 = fieldWeight in 2659, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.03125 = fieldNorm(doc=2659)
          0.024419045 = weight(_text_:22 in 2659) [ClassicSimilarity], result of:
            0.024419045 = score(doc=2659,freq=2.0), product of:
              0.15778607 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04505818 = queryNorm
              0.15476047 = fieldWeight in 2659, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=2659)
      0.4 = coord(2/5)
    
    Abstract
    The Library of Congress Subject Headings (LCSH) has been developed over the course of more than a century, predating the semantic web by some time. Until the 1986, the only concept-toconcept relationship available was an undifferentiated "See Also" reference, which was used for both associative (RT) and hierarchical (BT/NT) connections. In that year, in preparation for the first release of the headings in machine readable MARC Authorities form, an attempt was made to automatically convert these "See Also" links into the standardized thesaural relations. Unfortunately, the rule used to determine the type of reference to generate relied on the presence of symmetric links to detect associatively related terms; "See Also" references that were only present in one of the related terms were assumed to be hierarchical. This left the process vulnerable to inconsistent use of references in the pre-conversion data, with a marked bias towards promoting relationships to hierarchical status. The Library of Congress was aware that the results of the conversion contained many inconsistencies, and intended to validate and correct the results over the course of time. Unfortunately, twenty years later, less than 40% of the converted records have been evaluated. The converted records, being the earliest encountered during the Library's cataloging activities, represent the most basic concepts within LCSH; errors in the syndetic structure for these records affect far more subordinate concepts than those nearer the periphery. Worse, a policy of patterning new headings after pre-existing ones leads to structural errors arising from the conversion process being replicated in these newer headings, perpetuating and exacerbating the errors. As the LCSH prepares for its second great conversion, from MARC to SKOS, it is critical to address these structural problems. As part of the work on converting the headings into SKOS, I have experimented with different visualizations of the tangled web of broader terms embedded in LCSH. This poster illustrates several of these renderings, shows how they can help users to judge which relationships might not be correct, and shows just exactly how Doorbells and Mammals are related.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
  4. Mercun, T.; Zumer, M.; Aalberg, T.: Presenting bibliographic families using information visualization : evaluation of FRBR-based prototype and hierarchical visualizations (2017) 0.04
    0.044001963 = product of:
      0.1100049 = sum of:
        0.09240658 = weight(_text_:bibliographic in 3350) [ClassicSimilarity], result of:
          0.09240658 = score(doc=3350,freq=12.0), product of:
            0.17541347 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.04505818 = queryNorm
            0.52679294 = fieldWeight in 3350, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3350)
        0.017598324 = product of:
          0.035196647 = sum of:
            0.035196647 = weight(_text_:data in 3350) [ClassicSimilarity], result of:
              0.035196647 = score(doc=3350,freq=4.0), product of:
                0.14247625 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.04505818 = queryNorm
                0.24703519 = fieldWeight in 3350, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3350)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Since their beginnings, bibliographic information systems have been displaying results in the form of long, textual lists. With the development of new data models and computer technologies, the need for new approaches to present and interact with bibliographic data has slowly been maturing. To investigate how this could be accomplished, a prototype system, FrbrVis1, was designed to present work families within a bibliographic information system using information visualization. This paper reports on two user studies, a controlled and an observational experiment, that have been carried out to assess the Functional Requirements for Bibliographic Records (FRBR)-based against an existing system as well as to test four different hierarchical visual layouts. The results clearly show that FrbrVis offers better performance and user experience compared to the baseline system. The differences between the four hierarchical visualizations (Indented tree, Radial tree, Circlepack, and Sunburst) were, on the other hand, not as pronounced, but the Indented tree and Sunburst design proved to be the most successful, both in performance as well as user perception. The paper therefore not only evaluates the application of a visual presentation of bibliographic work families, but also provides valuable results regarding the performance and user acceptance of individual hierarchical visualization techniques.
  5. Mercun, T.; Zumer, M.; Aalberg, T.: Presenting bibliographic families : Designing an FRBR-based prototype using information visualization (2016) 0.04
    0.040781446 = product of:
      0.10195361 = sum of:
        0.08435529 = weight(_text_:bibliographic in 2879) [ClassicSimilarity], result of:
          0.08435529 = score(doc=2879,freq=10.0), product of:
            0.17541347 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.04505818 = queryNorm
            0.480894 = fieldWeight in 2879, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2879)
        0.017598324 = product of:
          0.035196647 = sum of:
            0.035196647 = weight(_text_:data in 2879) [ClassicSimilarity], result of:
              0.035196647 = score(doc=2879,freq=4.0), product of:
                0.14247625 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.04505818 = queryNorm
                0.24703519 = fieldWeight in 2879, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2879)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Purpose - Despite the importance of bibliographic information systems for discovering and exploring library resources, some of the core functionality that should be provided to support users in their information seeking process is still missing. Investigating these issues, the purpose of this paper is to design a solution that would fulfil the missing objectives. Design/methodology/approach - Building on the concepts of a work family, functional requirements for bibliographic records (FRBR) and information visualization, the paper proposes a model and user interface design that could support a more efficient and user-friendly presentation and navigation in bibliographic information systems. Findings - The proposed design brings together all versions of a work, related works, and other works by and about the author and shows how the model was implemented into a FrbrVis prototype system using hierarchical visualization layout. Research limitations/implications - Although issues related to discovery and exploration apply to various material types, the research first focused on works of fiction and was also limited by the selected sample of records. Practical implications - The model for presenting and interacting with FRBR-based data can serve as a good starting point for future developments and implementations. Originality/value - With FRBR concepts being gradually integrated into cataloguing rules, formats, and various bibliographic services, one of the important questions that has not really been investigated and studied is how the new type of data would be presented to users in a way that would exploit the true potential of the changes.
  6. Osinska, V.; Kowalska, M.; Osinski, Z.: ¬The role of visualization in the shaping and exploration of the individual information space : part 1 (2018) 0.04
    0.037254572 = product of:
      0.09313643 = sum of:
        0.03772483 = weight(_text_:bibliographic in 4641) [ClassicSimilarity], result of:
          0.03772483 = score(doc=4641,freq=2.0), product of:
            0.17541347 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.04505818 = queryNorm
            0.21506234 = fieldWeight in 4641, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4641)
        0.055411596 = sum of:
          0.024887787 = weight(_text_:data in 4641) [ClassicSimilarity], result of:
            0.024887787 = score(doc=4641,freq=2.0), product of:
              0.14247625 = queryWeight, product of:
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.04505818 = queryNorm
              0.17468026 = fieldWeight in 4641, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.0390625 = fieldNorm(doc=4641)
          0.030523809 = weight(_text_:22 in 4641) [ClassicSimilarity], result of:
            0.030523809 = score(doc=4641,freq=2.0), product of:
              0.15778607 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04505818 = queryNorm
              0.19345059 = fieldWeight in 4641, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=4641)
      0.4 = coord(2/5)
    
    Abstract
    Studies on the state and structure of digital knowledge concerning science generally relate to macro and meso scales. Supported by visualizations, these studies can deliver knowledge about emerging scientific fields or collaboration between countries, scientific centers, or groups of researchers. Analyses of individual activities or single scientific career paths are rarely presented and discussed. The authors decided to fill this gap and developed a web application for visualizing the scientific output of particular researchers. This free software based on bibliographic data from local databases, provides six layouts for analysis. Researchers can see the dynamic characteristics of their own writing activity, the time and place of publication, and the thematic scope of research problems. They can also identify cooperation networks, and consequently, study the dependencies and regularities in their own scientific activity. The current article presents the results of a study of the application's usability and functionality as well as attempts to define different user groups. A survey about the interface was sent to select researchers employed at Nicolaus Copernicus University. The results were used to answer the question as to whether such a specialized visualization tool can significantly augment the individual information space of the contemporary researcher.
    Date
    21.12.2018 17:22:13
  7. Platis, N. et al.: Visualization of uncertainty in tag clouds (2016) 0.02
    0.02216464 = product of:
      0.11082319 = sum of:
        0.11082319 = sum of:
          0.049775574 = weight(_text_:data in 2755) [ClassicSimilarity], result of:
            0.049775574 = score(doc=2755,freq=2.0), product of:
              0.14247625 = queryWeight, product of:
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.04505818 = queryNorm
              0.34936053 = fieldWeight in 2755, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.078125 = fieldNorm(doc=2755)
          0.061047617 = weight(_text_:22 in 2755) [ClassicSimilarity], result of:
            0.061047617 = score(doc=2755,freq=2.0), product of:
              0.15778607 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04505818 = queryNorm
              0.38690117 = fieldWeight in 2755, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.078125 = fieldNorm(doc=2755)
      0.2 = coord(1/5)
    
    Date
    1. 2.2016 18:25:22
    Source
    Semantic keyword-based search on structured data sources: First COST Action IC1302 International KEYSTONE Conference, IKC 2015, Coimbra, Portugal, September 8-9, 2015. Revised Selected Papers. Eds.: J. Cardoso et al
  8. Zhu, Y.; Yan, E.; Song, I.-Y..: ¬The use of a graph-based system to improve bibliographic information retrieval : system design, implementation, and evaluation (2017) 0.02
    0.020245269 = product of:
      0.101226345 = sum of:
        0.101226345 = weight(_text_:bibliographic in 3356) [ClassicSimilarity], result of:
          0.101226345 = score(doc=3356,freq=10.0), product of:
            0.17541347 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.04505818 = queryNorm
            0.5770728 = fieldWeight in 3356, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.046875 = fieldNorm(doc=3356)
      0.2 = coord(1/5)
    
    Abstract
    In this article, we propose a graph-based interactive bibliographic information retrieval system-GIBIR. GIBIR provides an effective way to retrieve bibliographic information. The system represents bibliographic information as networks and provides a form-based query interface. Users can develop their queries interactively by referencing the system-generated graph queries. Complex queries such as "papers on information retrieval, which were cited by John's papers that had been presented in SIGIR" can be effectively answered by the system. We evaluate the proposed system by developing another relational database-based bibliographic information retrieval system with the same interface and functions. Experiment results show that the proposed system executes the same queries much faster than the relational database-based system, and on average, our system reduced the execution time by 72% (for 3-node query), 89% (for 4-node query), and 99% (for 5-node query).
  9. Wen, B.; Horlings, E.; Zouwen, M. van der; Besselaar, P. van den: Mapping science through bibliometric triangulation : an experimental approach applied to water research (2017) 0.02
    0.020067489 = product of:
      0.050168723 = sum of:
        0.03772483 = weight(_text_:bibliographic in 3437) [ClassicSimilarity], result of:
          0.03772483 = score(doc=3437,freq=2.0), product of:
            0.17541347 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.04505818 = queryNorm
            0.21506234 = fieldWeight in 3437, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3437)
        0.012443894 = product of:
          0.024887787 = sum of:
            0.024887787 = weight(_text_:data in 3437) [ClassicSimilarity], result of:
              0.024887787 = score(doc=3437,freq=2.0), product of:
                0.14247625 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.04505818 = queryNorm
                0.17468026 = fieldWeight in 3437, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3437)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    The idea of constructing science maps based on bibliographic data has intrigued researchers for decades, and various techniques have been developed to map the structure of research disciplines. Most science mapping studies use a single method. However, as research fields have various properties, a valid map of a field should actually be composed of a set of maps derived from a series of investigations using different methods. That leads to the question of what can be learned from a combination-triangulation-of these different science maps. In this paper we propose a method for triangulation, using the example of water science. We combine three different mapping approaches: journal-journal citation relations (JJCR), shared author keywords (SAK), and title word-cited reference co-occurrence (TWRC). Our results demonstrate that triangulation of JJCR, SAK, and TWRC produces a more comprehensive picture than each method applied individually. The outcomes from the three different approaches can be associated with each other and systematically interpreted to provide insights into the complex multidisciplinary structure of the field of water research.
  10. Börner, K.: Atlas of knowledge : anyone can map (2015) 0.02
    0.01880732 = product of:
      0.094036594 = sum of:
        0.094036594 = sum of:
          0.042235978 = weight(_text_:data in 3355) [ClassicSimilarity], result of:
            0.042235978 = score(doc=3355,freq=4.0), product of:
              0.14247625 = queryWeight, product of:
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.04505818 = queryNorm
              0.29644224 = fieldWeight in 3355, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.046875 = fieldNorm(doc=3355)
          0.05180062 = weight(_text_:22 in 3355) [ClassicSimilarity], result of:
            0.05180062 = score(doc=3355,freq=4.0), product of:
              0.15778607 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04505818 = queryNorm
              0.32829654 = fieldWeight in 3355, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=3355)
      0.2 = coord(1/5)
    
    Date
    22. 1.2017 16:54:03
    22. 1.2017 17:10:56
    LCSH
    Communication in science / Data processing
    Subject
    Communication in science / Data processing
  11. Pejtersen, A.M.: Implications of users' value perception for the design of a bibliographic retrieval system (1986) 0.02
    0.018107919 = product of:
      0.09053959 = sum of:
        0.09053959 = weight(_text_:bibliographic in 2961) [ClassicSimilarity], result of:
          0.09053959 = score(doc=2961,freq=2.0), product of:
            0.17541347 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.04505818 = queryNorm
            0.5161496 = fieldWeight in 2961, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.09375 = fieldNorm(doc=2961)
      0.2 = coord(1/5)
    
  12. Börner, K.; Chen, C.; Boyack, K.W.: Visualizing knowledge domains (2002) 0.02
    0.015490483 = product of:
      0.038726207 = sum of:
        0.026407382 = weight(_text_:bibliographic in 4286) [ClassicSimilarity], result of:
          0.026407382 = score(doc=4286,freq=2.0), product of:
            0.17541347 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.04505818 = queryNorm
            0.15054363 = fieldWeight in 4286, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4286)
        0.012318826 = product of:
          0.024637653 = sum of:
            0.024637653 = weight(_text_:data in 4286) [ClassicSimilarity], result of:
              0.024637653 = score(doc=4286,freq=4.0), product of:
                0.14247625 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.04505818 = queryNorm
                0.17292464 = fieldWeight in 4286, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=4286)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    This chapter reviews visualization techniques that can be used to map the ever-growing domain structure of scientific disciplines and to support information retrieval and classification. In contrast to the comprehensive surveys conducted in traditional fashion by Howard White and Katherine McCain (1997, 1998), this survey not only reviews emerging techniques in interactive data analysis and information visualization, but also depicts the bibliographical structure of the field itself. The chapter starts by reviewing the history of knowledge domain visualization. We then present a general process flow for the visualization of knowledge domains and explain commonly used techniques. In order to visualize the domain reviewed by this chapter, we introduce a bibliographic data set of considerable size, which includes articles from the citation analysis, bibliometrics, semantics, and visualization literatures. Using tutorial style, we then apply various algorithms to demonstrate the visualization effectsl produced by different approaches and compare the results. The domain visualizations reveal the relationships within and between the four fields that together constitute the focus of this chapter. We conclude with a general discussion of research possibilities. Painting a "big picture" of scientific knowledge has long been desirable for a variety of reasons. Traditional approaches are brute forcescholars must sort through mountains of literature to perceive the outlines of their field. Obviously, this is time-consuming, difficult to replicate, and entails subjective judgments. The task is enormously complex. Sifting through recently published documents to find those that will later be recognized as important is labor intensive. Traditional approaches struggle to keep up with the pace of information growth. In multidisciplinary fields of study it is especially difficult to maintain an overview of literature dynamics. Painting the big picture of an everevolving scientific discipline is akin to the situation described in the widely known Indian legend about the blind men and the elephant. As the story goes, six blind men were trying to find out what an elephant looked like. They touched different parts of the elephant and quickly jumped to their conclusions. The one touching the body said it must be like a wall; the one touching the tail said it was like a snake; the one touching the legs said it was like a tree trunk, and so forth. But science does not stand still; the steady stream of new scientific literature creates a continuously changing structure. The resulting disappearance, fusion, and emergence of research areas add another twist to the tale-it is as if the elephant is running and dynamically changing its shape. Domain visualization, an emerging field of study, is in a similar situation. Relevant literature is spread across disciplines that have traditionally had few connections. Researchers examining the domain from a particular discipline cannot possibly have an adequate understanding of the whole. As noted by White and McCain (1997), the new generation of information scientists is technically driven in its efforts to visualize scientific disciplines. However, limited progress has been made in terms of connecting pioneers' theories and practices with the potentialities of today's enabling technologies. If the difference between past and present generations lies in the power of available technologies, what they have in common is the ultimate goal-to reveal the development of scientific knowledge.
  13. Di Maio, P.: Linked data beyond libraries : towards universal interfaces and knowledge unification (2015) 0.01
    0.008447195 = product of:
      0.042235978 = sum of:
        0.042235978 = product of:
          0.084471956 = sum of:
            0.084471956 = weight(_text_:data in 2553) [ClassicSimilarity], result of:
              0.084471956 = score(doc=2553,freq=4.0), product of:
                0.14247625 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.04505818 = queryNorm
                0.5928845 = fieldWeight in 2553, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.09375 = fieldNorm(doc=2553)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Source
    Linked data and user interaction: the road ahead. Eds.: Cervone, H.F. u. L.G. Svensson
  14. Lamb, I.; Larson, C.: Shining a light on scientific data : building a data catalog to foster data sharing and reuse (2016) 0.01
    0.008447195 = product of:
      0.042235978 = sum of:
        0.042235978 = product of:
          0.084471956 = sum of:
            0.084471956 = weight(_text_:data in 3195) [ClassicSimilarity], result of:
              0.084471956 = score(doc=3195,freq=16.0), product of:
                0.14247625 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.04505818 = queryNorm
                0.5928845 = fieldWeight in 3195, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3195)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Abstract
    The scientific community's growing eagerness to make research data available to the public provides libraries - with our expertise in metadata and discovery - an interesting new opportunity. This paper details the in-house creation of a "data catalog" which describes datasets ranging from population-level studies like the US Census to small, specialized datasets created by researchers at our own institution. Based on Symfony2 and Solr, the data catalog provides a powerful search interface to help researchers locate the data that can help them, and an administrative interface so librarians can add, edit, and manage metadata elements at will. This paper will outline the successes, failures, and total redos that culminated in the current manifestation of our data catalog.
  15. Bornmann, L.; Haunschild, R.: Overlay maps based on Mendeley data : the use of altmetrics for readership networks (2016) 0.01
    0.0079016285 = product of:
      0.03950814 = sum of:
        0.03950814 = product of:
          0.07901628 = sum of:
            0.07901628 = weight(_text_:data in 3230) [ClassicSimilarity], result of:
              0.07901628 = score(doc=3230,freq=14.0), product of:
                0.14247625 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.04505818 = queryNorm
                0.55459267 = fieldWeight in 3230, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3230)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Abstract
    Visualization of scientific results using networks has become popular in scientometric research. We provide base maps for Mendeley reader count data using the publication year 2012 from the Web of Science data. Example networks are shown and explained. The reader can use our base maps to visualize other results with the VOSViewer. The proposed overlay maps are able to show the impact of publications in terms of readership data. The advantage of using our base maps is that it is not necessary for the user to produce a network based on all data (e.g., from 1 year), but can collect the Mendeley data for a single institution (or journals, topics) and can match them with our already produced information. Generation of such large-scale networks is still a demanding task despite the available computer power and digital data availability. Therefore, it is very useful to have base maps and create the network with the overlay technique.
  16. Rohner, M.: Betrachtung der Data Visualization Literacy in der angestrebten Schweizer Informationsgesellschaft (2018) 0.01
    0.0079016285 = product of:
      0.03950814 = sum of:
        0.03950814 = product of:
          0.07901628 = sum of:
            0.07901628 = weight(_text_:data in 4585) [ClassicSimilarity], result of:
              0.07901628 = score(doc=4585,freq=14.0), product of:
                0.14247625 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.04505818 = queryNorm
                0.55459267 = fieldWeight in 4585, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4585)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Abstract
    Datenvisualisierungen sind ein wichtiges Werkzeug, um Inhalte und Muster in Datensätzen zu erkennen und ermöglichen so auch Laien den Zugang zu der Information, die in Datensätzen steckt. Data Visualization Literacy ist die Kompetenz, Datenvisualisierungen zu lesen, zu verstehen, zu hinterfragen und herzustellen. Data Visulaization Literacy ist daher eine wichtige Kompetenz der Informationsgesellschaft. Im Auftrag des Bundesrates hat das Bundesamt für Kommunikation BAKOM die Strategie "Digitale Schweiz" entwickelt. Die Strategie zeigt auf, wie die fortschreitende Digitalisierung genutzt und die Schweiz zu einer Informationsgesellschaft entwickelt werden soll. In der vorliegenden Arbeit wird untersucht, inwiefern die Strategie "Digitale Schweiz" die Förderung von Data Visualization Literacy in der Bevölkerung unterstützt. Dazu werden die Kompetenzen der Data Visualization Literacy ermittelt, Kompetenzstellen innerhalb des Bildungssystems benannt und die Massnahmen der Strategie in Bezug auf Data Visualization Literacy überprüft.
    Content
    Diese Publikation entstand im Rahmen einer Thesis zum Master of Science FHO in Business Administration, Major Information and Data Management.
  17. Tang, M.-C.: Browsing and searching in a faceted information space : a naturalistic study of PubMed users' interaction with a display tool (2007) 0.01
    0.007544966 = product of:
      0.03772483 = sum of:
        0.03772483 = weight(_text_:bibliographic in 617) [ClassicSimilarity], result of:
          0.03772483 = score(doc=617,freq=2.0), product of:
            0.17541347 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.04505818 = queryNorm
            0.21506234 = fieldWeight in 617, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.0390625 = fieldNorm(doc=617)
      0.2 = coord(1/5)
    
    Abstract
    The study adopts a naturalistic approach to investigate users' interaction with a browsable MeSH (medical subject headings) display designed to facilitate query construction for the PubMed bibliographic database. The purpose of the study is twofold: first, to test the usefulness of a browsable interface utilizing the principle of faceted classification; and second, to investigate users' preferred query submission methods in different problematic situations. An interface that incorporated multiple query submission methods - the conventional single-line query box as well as methods associated the faceted classification display was constructed. Participants' interactions with the interface were monitored remotely over a period of 10 weeks; information about their problematic situations and information retrieval behaviors were also collected during this time. The traditional controlled experiment was not adequate in answering the author's research questions; hence, the author provides his rationale for a naturalistic approach. The study's findings show that there is indeed a selective compatibility between query submission methods provided by the MeSH display and users' problematic situations. The query submission methods associated with the display were found to be the preferred search tools when users' information needs were vague and the search topics unfamiliar. The findings support the theoretical proposition that users engaging in an information retrieval process with a variety of problematic situations need different approaches. The author argues that rather than treat the information retrieval system as a general purpose tool, more attention should be given to the interaction between the functionality of the tool and the characteristics of users' problematic situations.
  18. Choi, I.: Visualizations of cross-cultural bibliographic classification : comparative studies of the Korean Decimal Classification and the Dewey Decimal Classification (2017) 0.01
    0.007544966 = product of:
      0.03772483 = sum of:
        0.03772483 = weight(_text_:bibliographic in 3869) [ClassicSimilarity], result of:
          0.03772483 = score(doc=3869,freq=2.0), product of:
            0.17541347 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.04505818 = queryNorm
            0.21506234 = fieldWeight in 3869, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3869)
      0.2 = coord(1/5)
    
  19. Maaten, L. van den; Hinton, G.: Visualizing data using t-SNE (2008) 0.01
    0.0074663362 = product of:
      0.03733168 = sum of:
        0.03733168 = product of:
          0.07466336 = sum of:
            0.07466336 = weight(_text_:data in 3888) [ClassicSimilarity], result of:
              0.07466336 = score(doc=3888,freq=18.0), product of:
                0.14247625 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.04505818 = queryNorm
                0.52404076 = fieldWeight in 3888, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3888)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Abstract
    We present a new technique called "t-SNE" that visualizes high-dimensional data by giving each datapoint a location in a two or three-dimensional map. The technique is a variation of Stochastic Neighbor Embedding (Hinton and Roweis, 2002) that is much easier to optimize, and produces significantly better visualizations by reducing the tendency to crowd points together in the center of the map. t-SNE is better than existing techniques at creating a single map that reveals structure at many different scales. This is particularly important for high-dimensional data that lie on several different, but related, low-dimensional manifolds, such as images of objects from multiple classes seen from multiple viewpoints. For visualizing the structure of very large data sets, we show how t-SNE can use random walks on neighborhood graphs to allow the implicit structure of all of the data to influence the way in which a subset of the data is displayed. We illustrate the performance of t-SNE on a wide variety of data sets and compare it with many other non-parametric visualization techniques, including Sammon mapping, Isomap, and Locally Linear Embedding. The visualizations produced by t-SNE are significantly better than those produced by the other techniques on almost all of the data sets.
    Theme
    Data Mining
  20. Maaten, L. van den; Hinton, G.: Visualizing non-metric similarities in multiple maps (2012) 0.01
    0.007315486 = product of:
      0.03657743 = sum of:
        0.03657743 = product of:
          0.07315486 = sum of:
            0.07315486 = weight(_text_:data in 3884) [ClassicSimilarity], result of:
              0.07315486 = score(doc=3884,freq=12.0), product of:
                0.14247625 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.04505818 = queryNorm
                0.513453 = fieldWeight in 3884, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3884)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Abstract
    Techniques for multidimensional scaling visualize objects as points in a low-dimensional metric map. As a result, the visualizations are subject to the fundamental limitations of metric spaces. These limitations prevent multidimensional scaling from faithfully representing non-metric similarity data such as word associations or event co-occurrences. In particular, multidimensional scaling cannot faithfully represent intransitive pairwise similarities in a visualization, and it cannot faithfully visualize "central" objects. In this paper, we present an extension of a recently proposed multidimensional scaling technique called t-SNE. The extension aims to address the problems of traditional multidimensional scaling techniques when these techniques are used to visualize non-metric similarities. The new technique, called multiple maps t-SNE, alleviates these problems by constructing a collection of maps that reveal complementary structure in the similarity data. We apply multiple maps t-SNE to a large data set of word association data and to a data set of NIPS co-authorships, demonstrating its ability to successfully visualize non-metric similarities.
    Theme
    Data Mining

Years

Languages

  • e 68
  • d 8
  • a 1
  • More… Less…

Types

  • a 63
  • el 21
  • m 6
  • x 4
  • r 1
  • s 1
  • More… Less…