Search (130 results, page 1 of 7)

  • × theme_ss:"Visualisierung"
  • × type_ss:"a"
  1. Zhu, B.; Chen, H.: Information visualization (2004) 0.05
    0.050138313 = product of:
      0.1128112 = sum of:
        0.029363085 = weight(_text_:applications in 4276) [ClassicSimilarity], result of:
          0.029363085 = score(doc=4276,freq=2.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.17024462 = fieldWeight in 4276, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4276)
        0.02191663 = weight(_text_:of in 4276) [ClassicSimilarity], result of:
          0.02191663 = score(doc=4276,freq=70.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.35774738 = fieldWeight in 4276, product of:
              8.3666 = tf(freq=70.0), with freq of:
                70.0 = termFreq=70.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4276)
        0.020234404 = weight(_text_:systems in 4276) [ClassicSimilarity], result of:
          0.020234404 = score(doc=4276,freq=4.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.16806422 = fieldWeight in 4276, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4276)
        0.041297078 = weight(_text_:software in 4276) [ClassicSimilarity], result of:
          0.041297078 = score(doc=4276,freq=6.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.26571283 = fieldWeight in 4276, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4276)
      0.44444445 = coord(4/9)
    
    Abstract
    Advanced technology has resulted in the generation of about one million terabytes of information every year. Ninety-reine percent of this is available in digital format (Keim, 2001). More information will be generated in the next three years than was created during all of previous human history (Keim, 2001). Collecting information is no longer a problem, but extracting value from information collections has become progressively more difficult. Various search engines have been developed to make it easier to locate information of interest, but these work well only for a person who has a specific goal and who understands what and how information is stored. This usually is not the Gase. Visualization was commonly thought of in terms of representing human mental processes (MacEachren, 1991; Miller, 1984). The concept is now associated with the amplification of these mental processes (Card, Mackinlay, & Shneiderman, 1999). Human eyes can process visual cues rapidly, whereas advanced information analysis techniques transform the computer into a powerful means of managing digitized information. Visualization offers a link between these two potent systems, the human eye and the computer (Gershon, Eick, & Card, 1998), helping to identify patterns and to extract insights from large amounts of information. The identification of patterns is important because it may lead to a scientific discovery, an interpretation of clues to solve a crime, the prediction of catastrophic weather, a successful financial investment, or a better understanding of human behavior in a computermediated environment. Visualization technology shows considerable promise for increasing the value of large-scale collections of information, as evidenced by several commercial applications of TreeMap (e.g., http://www.smartmoney.com) and Hyperbolic tree (e.g., http://www.inxight.com) to visualize large-scale hierarchical structures. Although the proliferation of visualization technologies dates from the 1990s where sophisticated hardware and software made increasingly faster generation of graphical objects possible, the role of visual aids in facilitating the construction of mental images has a long history. Visualization has been used to communicate ideas, to monitor trends implicit in data, and to explore large volumes of data for hypothesis generation. Imagine traveling to a strange place without a map, having to memorize physical and chemical properties of an element without Mendeleyev's periodic table, trying to understand the stock market without statistical diagrams, or browsing a collection of documents without interactive visual aids. A collection of information can lose its value simply because of the effort required for exhaustive exploration. Such frustrations can be overcome by visualization.
    Visualization can be classified as scientific visualization, software visualization, or information visualization. Although the data differ, the underlying techniques have much in common. They use the same elements (visual cues) and follow the same rules of combining visual cues to deliver patterns. They all involve understanding human perception (Encarnacao, Foley, Bryson, & Feiner, 1994) and require domain knowledge (Tufte, 1990). Because most decisions are based an unstructured information, such as text documents, Web pages, or e-mail messages, this chapter focuses an the visualization of unstructured textual documents. The chapter reviews information visualization techniques developed over the last decade and examines how they have been applied in different domains. The first section provides the background by describing visualization history and giving overviews of scientific, software, and information visualization as well as the perceptual aspects of visualization. The next section assesses important visualization techniques that convert abstract information into visual objects and facilitate navigation through displays an a computer screen. It also explores information analysis algorithms that can be applied to identify or extract salient visualizable structures from collections of information. Information visualization systems that integrate different types of technologies to address problems in different domains are then surveyed; and we move an to a survey and critique of visualization system evaluation studies. The chapter concludes with a summary and identification of future research directions.
    Source
    Annual review of information science and technology. 39(2005), S.139-177
  2. Wu, I.-C.; Vakkari, P.: Supporting navigation in Wikipedia by information visualization : extended evaluation measures (2014) 0.03
    0.030291064 = product of:
      0.09087319 = sum of:
        0.05812384 = weight(_text_:applications in 1797) [ClassicSimilarity], result of:
          0.05812384 = score(doc=1797,freq=6.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.33699697 = fieldWeight in 1797, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03125 = fieldNorm(doc=1797)
        0.016397487 = weight(_text_:of in 1797) [ClassicSimilarity], result of:
          0.016397487 = score(doc=1797,freq=30.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.26765788 = fieldWeight in 1797, product of:
              5.477226 = tf(freq=30.0), with freq of:
                30.0 = termFreq=30.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=1797)
        0.016351866 = weight(_text_:systems in 1797) [ClassicSimilarity], result of:
          0.016351866 = score(doc=1797,freq=2.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.1358164 = fieldWeight in 1797, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03125 = fieldNorm(doc=1797)
      0.33333334 = coord(3/9)
    
    Abstract
    Purpose - The authors introduce two semantics-based navigation applications that facilitate information-seeking activities in internal link-based web sites in Wikipedia. These applications aim to help users find concepts within a topic and related articles on a given topic quickly and then gain topical knowledge from internal link-based encyclopedia web sites. The paper aims to discuss these issues. Design/methodology/approach - The WNavis application consists of three information visualization (IV) tools which are a topic network, a hierarchy topic tree and summaries for topics. The WikiMap application consists of a topic network. The goal of the topic network and topic tree tools is to help users to find the major concepts of a topic and identify relationships between these major concepts easily. In addition, in order to locate specific information and enable users to explore and read topic-related articles quickly, the topic tree and summaries for topics tools support users to gain topical knowledge quickly. The authors then apply the k-clique of cohesive indicator to analyze the sub topics of the seed query and find out the best clustering results via the cosine measure. The authors utilize four metrics, which are correctness, time cost, usage behaviors, and satisfaction, to evaluate the three interfaces. These metrics measure both the outputs and outcomes of applications. As a baseline system for evaluation the authors used a traditional Wikipedia interface. For the evaluation, the authors used an experimental user study with 30 participants.
    Findings - The results indicate that both WikiMap and WNavis supported users to identify concepts and their relations better compared to the baseline. In topical tasks WNavis over performed both WikiMap and the baseline system. Although there were no time differences in finding concepts or answering topical questions, the test systems provided users with a greater gain per time unit. The users of WNavis leaned on the hierarchy tree instead of other tools, whereas WikiMap users used the topic map. Research limitations/implications - The findings have implications for the design of IR support tools in knowledge-intensive web sites that help users to explore topics and concepts. Originality/value - The authors explored to what extent the use of each IV support tool contributed to successful exploration of topics in search tasks. The authors propose extended task-based evaluation measures to understand how each application provides useful context for users to accomplish the tasks and attain the search goals. That is, the authors not only evaluate the output of the search results, e.g. the number of relevant items retrieved, but also the outcome provided by the system for assisting users to attain the search goal.
    Source
    Journal of documentation. 70(2014) no.3, S.392-424
  3. Eito Brun, R.: Retrieval effectiveness in software repositories : from faceted classifications to software visualization techniques (2006) 0.03
    0.029939393 = product of:
      0.13472727 = sum of:
        0.016735615 = weight(_text_:of in 2515) [ClassicSimilarity], result of:
          0.016735615 = score(doc=2515,freq=20.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.27317715 = fieldWeight in 2515, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2515)
        0.117991656 = weight(_text_:software in 2515) [ClassicSimilarity], result of:
          0.117991656 = score(doc=2515,freq=24.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.75917953 = fieldWeight in 2515, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2515)
      0.22222222 = coord(2/9)
    
    Abstract
    The internal organization of large software projects requires an extraordinary effort in the development and maintenance of repositories made up of software artifacts (business components, data models, functional and technical documentation, etc.). During the software development process, different artifacts are created to help users in the transfer of knowledge and enable communication between workers and teams. The storage, maintenance and publication of these artifacts in knowledge bases - usually referred to as "software repositories" are a useful tool for future software development projects, as they contain the collective, learned experience of the teams and provide the basis to estimate and reuse the work completed in the past. Different techniques similar to those used by the library community have been used in the past to organize these software repositories and help users in the difficult task or identifying and retrieving artifacts (software and documentation). These techniques include software classification - with a special emphasis on faceted classifications, keyword-based retrieval and formal method techniques. The paper discusses the different knowledge organization techniques applied in these repositories to identify and retrieve software artifacts and ensure the reusability of software components and documentation at the different phases of the development process across different projects. An enumeration of the main approaches documented in specialized bibliography is provided.
    Source
    Knowledge organization for a global learning society: Proceedings of the 9th International ISKO Conference, 4-7 July 2006, Vienna, Austria. Hrsg.: G. Budin, C. Swertz u. K. Mitgutsch
  4. Zhang, J.; Nguyen, T.: WebStar: a visualization model for hyperlink structures (2005) 0.03
    0.029688384 = product of:
      0.08906515 = sum of:
        0.050336715 = weight(_text_:applications in 1056) [ClassicSimilarity], result of:
          0.050336715 = score(doc=1056,freq=2.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.2918479 = fieldWeight in 1056, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.046875 = fieldNorm(doc=1056)
        0.014200641 = weight(_text_:of in 1056) [ClassicSimilarity], result of:
          0.014200641 = score(doc=1056,freq=10.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.23179851 = fieldWeight in 1056, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=1056)
        0.0245278 = weight(_text_:systems in 1056) [ClassicSimilarity], result of:
          0.0245278 = score(doc=1056,freq=2.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.2037246 = fieldWeight in 1056, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.046875 = fieldNorm(doc=1056)
      0.33333334 = coord(3/9)
    
    Abstract
    The authors introduce an information visualization model, WebStar, for hyperlink-based information systems. Hyperlinks within a hyperlink-based document can be visualized in a two-dimensional visual space. All links are projected within a display sphere in the visual space. The relationship between a specified central document and its hyperlinked documents is visually presented in the visual space. In addition, users are able to define a group of subjects and to observe relevance between each subject and all hyperlinked documents via movement of that subject around the display sphere center. WebStar allows users to dynamically change an interest center during navigation. A retrieval mechanism is developed to control retrieved results in the visual space. Impact of movement of a subject on the visual document distribution is analyzed. An ambiguity problem caused by projection is discussed. Potential applications of this visualization model in information retrieval are included. Future research directions on the topic are addressed.
  5. Spero, S.: LCSH is to thesaurus as doorbell is to mammal : visualizing structural problems in the Library of Congress Subject Headings (2008) 0.03
    0.02534542 = product of:
      0.07603626 = sum of:
        0.04745791 = weight(_text_:applications in 2659) [ClassicSimilarity], result of:
          0.04745791 = score(doc=2659,freq=4.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.27515686 = fieldWeight in 2659, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03125 = fieldNorm(doc=2659)
        0.017962547 = weight(_text_:of in 2659) [ClassicSimilarity], result of:
          0.017962547 = score(doc=2659,freq=36.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.2932045 = fieldWeight in 2659, product of:
              6.0 = tf(freq=36.0), with freq of:
                36.0 = termFreq=36.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=2659)
        0.010615807 = product of:
          0.021231614 = sum of:
            0.021231614 = weight(_text_:22 in 2659) [ClassicSimilarity], result of:
              0.021231614 = score(doc=2659,freq=2.0), product of:
                0.13719016 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03917671 = queryNorm
                0.15476047 = fieldWeight in 2659, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2659)
          0.5 = coord(1/2)
      0.33333334 = coord(3/9)
    
    Abstract
    The Library of Congress Subject Headings (LCSH) has been developed over the course of more than a century, predating the semantic web by some time. Until the 1986, the only concept-toconcept relationship available was an undifferentiated "See Also" reference, which was used for both associative (RT) and hierarchical (BT/NT) connections. In that year, in preparation for the first release of the headings in machine readable MARC Authorities form, an attempt was made to automatically convert these "See Also" links into the standardized thesaural relations. Unfortunately, the rule used to determine the type of reference to generate relied on the presence of symmetric links to detect associatively related terms; "See Also" references that were only present in one of the related terms were assumed to be hierarchical. This left the process vulnerable to inconsistent use of references in the pre-conversion data, with a marked bias towards promoting relationships to hierarchical status. The Library of Congress was aware that the results of the conversion contained many inconsistencies, and intended to validate and correct the results over the course of time. Unfortunately, twenty years later, less than 40% of the converted records have been evaluated. The converted records, being the earliest encountered during the Library's cataloging activities, represent the most basic concepts within LCSH; errors in the syndetic structure for these records affect far more subordinate concepts than those nearer the periphery. Worse, a policy of patterning new headings after pre-existing ones leads to structural errors arising from the conversion process being replicated in these newer headings, perpetuating and exacerbating the errors. As the LCSH prepares for its second great conversion, from MARC to SKOS, it is critical to address these structural problems. As part of the work on converting the headings into SKOS, I have experimented with different visualizations of the tangled web of broader terms embedded in LCSH. This poster illustrates several of these renderings, shows how they can help users to judge which relationships might not be correct, and shows just exactly how Doorbells and Mammals are related.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
  6. Cobo, M.J.; López-Herrera, A.G.; Herrera-Viedma, E.; Herrera, F.: Science mapping software tools : review, analysis, and cooperative study among tools (2011) 0.02
    0.024739683 = product of:
      0.11132857 = sum of:
        0.016935252 = weight(_text_:of in 4486) [ClassicSimilarity], result of:
          0.016935252 = score(doc=4486,freq=8.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.27643585 = fieldWeight in 4486, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=4486)
        0.09439332 = weight(_text_:software in 4486) [ClassicSimilarity], result of:
          0.09439332 = score(doc=4486,freq=6.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.6073436 = fieldWeight in 4486, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0625 = fieldNorm(doc=4486)
      0.22222222 = coord(2/9)
    
    Abstract
    Science mapping aims to build bibliometric maps that describe how specific disciplines, scientific domains, or research fields are conceptually, intellectually, and socially structured. Different techniques and software tools have been proposed to carry out science mapping analysis. The aim of this article is to review, analyze, and compare some of these software tools, taking into account aspects such as the bibliometric techniques available and the different kinds of analysis.
    Source
    Journal of the American Society for Information Science and Technology. 62(2011) no.7, S.1382-1402
  7. Catarci, T.; Spaccapietra, S.: Visual information querying (2002) 0.02
    0.02323037 = product of:
      0.06969111 = sum of:
        0.025168357 = weight(_text_:applications in 4268) [ClassicSimilarity], result of:
          0.025168357 = score(doc=4268,freq=2.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.14592396 = fieldWeight in 4268, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.0234375 = fieldNorm(doc=4268)
        0.017099835 = weight(_text_:of in 4268) [ClassicSimilarity], result of:
          0.017099835 = score(doc=4268,freq=58.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.27912235 = fieldWeight in 4268, product of:
              7.615773 = tf(freq=58.0), with freq of:
                58.0 = termFreq=58.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0234375 = fieldNorm(doc=4268)
        0.027422912 = weight(_text_:systems in 4268) [ClassicSimilarity], result of:
          0.027422912 = score(doc=4268,freq=10.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.22777101 = fieldWeight in 4268, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.0234375 = fieldNorm(doc=4268)
      0.33333334 = coord(3/9)
    
    Abstract
    Computers have become our companions in many of the activities we pursue in our life. They assist us, in particular, in searching relevant information that is needed to perform a variety of tasks, from professional usage to personal entertainment. They hold this information in a huge number of heterogeneous sources, either dedicated to a specific user community (e.g., enterprise databases) or maintained for the general public (e.g., websites and digital libraries). Whereas progress in basic information technology is nowadays capable of guaranteeing effective information management, information retrieval and dissemination has become a core issue that needs further accomplishments to achieve user satisfaction. The research communities in databases, information retrieval, information visualization, and human-computer interaction have already largely investigated these domains. However, the technical environment has so dramatically evolved in recent years, inducing a parallel and very significant evolution in user habits and expectations, that new approaches are definitely needed to meet current demand. One of the most evident and significant changes is the human-computer interaction paradigm. Traditional interactions relayed an programming to express user information requirements in formal code and an textual output to convey to users the information extracted by the system. Except for professional data-intensive application frameworks, still in the hands of computer speciahsts, we have basically moved away from this pattern both in terms of expressing information requests and conveying results. The new goal is direct interaction with the final user (the person who is looking for information and is not necessarily familiar with computer technology). The key motto to achieve this is "go visual." The well-known high bandwidth of the human-vision channel allows both recognition and understanding of large quantities of information in no more than a few seconds. Thus, for instance, if the result of an information request can be organized as a visual display, or a sequence of visual displays, the information throughput is immensely superior to the one that can be achieved using textual support. User interaction becomes an iterative query-answer game that very rapidly leads to the desired final result. Conversely, the system can provide efficient visual support for easy query formulation. Displaying a visual representation of the information space, for instance, lets users directly point at the information they are looking for, without any need to be trained into the complex syntax of current query languages. Alternatively, users can navigate in the information space, following visible paths that will lead them to the targeted items. Again, thanks to the visual support, users are able to easily understand how to formulate queries and they are likely to achieve the task more rapidly and less prone to errors than with traditional textual interaction modes.
    The two facets of "going visual" are usually referred to as visual query systems, for query formulation, and information visualization, for result display. Visual Query Systems (VQSs) are defined as systems for querying databases that use a visual representation to depict the domain of interest and express related requests. VQSs provide both a language to express the queries in a visual format and a variety of functionalities to facilitate user-system interaction. As such, they are oriented toward a wide spectrum of users, especially novices who have limited computer expertise and generally ignore the inner structure of the accessed database. Information visualization, an increasingly important subdiscipline within the field of Human-Computer Interaction (HCI), focuses an visual mechanisms designed to communicate clearly to the user the structure of information and improve an the cost of accessing large data repositories. In printed form, information visualization has included the display of numerical data (e.g., bar charts, plot charts, pie charts), combinatorial relations (e.g., drawings of graphs), and geographic data (e.g., encoded maps). In addition to these "static" displays, computer-based systems, such as the Information Visualizer and Dynamic Queries, have coupled powerful visualization techniques (e.g., 3D, animation) with near real-time interactivity (i.e., the ability of the system to respond quickly to the user's direct manipulation commands). Information visualization is tightly combined with querying capabilities in some recent database-centered approaches. More opportunities for information visualization in a database environment may be found today in data mining and data warehousing applications, which typically access large data repositories. The enormous quantity of information sources an the World-Wide Web (WWW) available to users with diverse capabilities also calls for visualization techniques. In this article, we survey the main features and main proposals for visual query systems and touch upon the visualization of results mainly discussing traditional visualization forms. A discussion of modern database visualization techniques may be found elsewhere. Many related articles by Daniel Keim are available at http://www. informatik.uni-halle.de/dbs/publications.html.
    Source
    Encyclopedia of library and information science. Vol.72, [=Suppl.35]
  8. Pejtersen, A.M.: Implications of users' value perception for the design of a bibliographic retrieval system (1986) 0.02
    0.023054788 = product of:
      0.10374655 = sum of:
        0.021999538 = weight(_text_:of in 2961) [ClassicSimilarity], result of:
          0.021999538 = score(doc=2961,freq=6.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.3591007 = fieldWeight in 2961, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.09375 = fieldNorm(doc=2961)
        0.08174701 = weight(_text_:software in 2961) [ClassicSimilarity], result of:
          0.08174701 = score(doc=2961,freq=2.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.525975 = fieldWeight in 2961, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.09375 = fieldNorm(doc=2961)
      0.22222222 = coord(2/9)
    
    Source
    Empirical foundation of information and software science. Ed.: J.C. Agarwal u. P. Zunde
  9. Osinska, V.; Kowalska, M.; Osinski, Z.: ¬The role of visualization in the shaping and exploration of the individual information space : part 1 (2018) 0.02
    0.021887988 = product of:
      0.06566396 = sum of:
        0.018332949 = weight(_text_:of in 4641) [ClassicSimilarity], result of:
          0.018332949 = score(doc=4641,freq=24.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.2992506 = fieldWeight in 4641, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4641)
        0.034061253 = weight(_text_:software in 4641) [ClassicSimilarity], result of:
          0.034061253 = score(doc=4641,freq=2.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.21915624 = fieldWeight in 4641, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4641)
        0.013269759 = product of:
          0.026539518 = sum of:
            0.026539518 = weight(_text_:22 in 4641) [ClassicSimilarity], result of:
              0.026539518 = score(doc=4641,freq=2.0), product of:
                0.13719016 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03917671 = queryNorm
                0.19345059 = fieldWeight in 4641, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4641)
          0.5 = coord(1/2)
      0.33333334 = coord(3/9)
    
    Abstract
    Studies on the state and structure of digital knowledge concerning science generally relate to macro and meso scales. Supported by visualizations, these studies can deliver knowledge about emerging scientific fields or collaboration between countries, scientific centers, or groups of researchers. Analyses of individual activities or single scientific career paths are rarely presented and discussed. The authors decided to fill this gap and developed a web application for visualizing the scientific output of particular researchers. This free software based on bibliographic data from local databases, provides six layouts for analysis. Researchers can see the dynamic characteristics of their own writing activity, the time and place of publication, and the thematic scope of research problems. They can also identify cooperation networks, and consequently, study the dependencies and regularities in their own scientific activity. The current article presents the results of a study of the application's usability and functionality as well as attempts to define different user groups. A survey about the interface was sent to select researchers employed at Nicolaus Copernicus University. The results were used to answer the question as to whether such a specialized visualization tool can significantly augment the individual information space of the contemporary researcher.
    Date
    21.12.2018 17:22:13
  10. Beagle, D.: Visualizing keyword distribution across multidisciplinary c-space (2003) 0.02
    0.018078228 = product of:
      0.054234684 = sum of:
        0.025168357 = weight(_text_:applications in 1202) [ClassicSimilarity], result of:
          0.025168357 = score(doc=1202,freq=2.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.14592396 = fieldWeight in 1202, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1202)
        0.016802425 = weight(_text_:of in 1202) [ClassicSimilarity], result of:
          0.016802425 = score(doc=1202,freq=56.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.2742677 = fieldWeight in 1202, product of:
              7.483315 = tf(freq=56.0), with freq of:
                56.0 = termFreq=56.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1202)
        0.0122639 = weight(_text_:systems in 1202) [ClassicSimilarity], result of:
          0.0122639 = score(doc=1202,freq=2.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.1018623 = fieldWeight in 1202, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1202)
      0.33333334 = coord(3/9)
    
    Abstract
    The concept of c-space is proposed as a visualization schema relating containers of content to cataloging surrogates and classification structures. Possible applications of keyword vector clusters within c-space could include improved retrieval rates through the use of captioning within visual hierarchies, tracings of semantic bleeding among subclasses, and access to buried knowledge within subject-neutral publication containers. The Scholastica Project is described as one example, following a tradition of research dating back to the 1980's. Preliminary focus group assessment indicates that this type of classification rendering may offer digital library searchers enriched entry strategies and an expanded range of re-entry vocabularies. Those of us who work in traditional libraries typically assume that our systems of classification: Library of Congress Classification (LCC) and Dewey Decimal Classification (DDC), are descriptive rather than prescriptive. In other words, LCC classes and subclasses approximate natural groupings of texts that reflect an underlying order of knowledge, rather than arbitrary categories prescribed by librarians to facilitate efficient shelving. Philosophical support for this assumption has traditionally been found in a number of places, from the archetypal tree of knowledge, to Aristotelian categories, to the concept of discursive formations proposed by Michel Foucault. Gary P. Radford has elegantly described an encounter with Foucault's discursive formations in the traditional library setting: "Just by looking at the titles on the spines, you can see how the books cluster together...You can identify those books that seem to form the heart of the discursive formation and those books that reside on the margins. Moving along the shelves, you see those books that tend to bleed over into other classifications and that straddle multiple discursive formations. You can physically and sensually experience...those points that feel like state borders or national boundaries, those points where one subject ends and another begins, or those magical places where one subject has morphed into another..."
    But what happens to this awareness in a digital library? Can discursive formations be represented in cyberspace, perhaps through diagrams in a visualization interface? And would such a schema be helpful to a digital library user? To approach this question, it is worth taking a moment to reconsider what Radford is looking at. First, he looks at titles to see how the books cluster. To illustrate, I scanned one hundred books on the shelves of a college library under subclass HT 101-395, defined by the LCC subclass caption as Urban groups. The City. Urban sociology. Of the first 100 titles in this sequence, fifty included the word "urban" or variants (e.g. "urbanization"). Another thirty-five used the word "city" or variants. These keywords appear to mark their titles as the heart of this discursive formation. The scattering of titles not using "urban" or "city" used related terms such as "town," "community," or in one case "skyscrapers." So we immediately see some empirical correlation between keywords and classification. But we also see a problem with the commonly used search technique of title-keyword. A student interested in urban studies will want to know about this entire subclass, and may wish to browse every title available therein. A title-keyword search on "urban" will retrieve only half of the titles, while a search on "city" will retrieve just over a third. There will be no overlap, since no titles in this sample contain both words. The only place where both words appear in a common string is in the LCC subclass caption, but captions are not typically indexed in library Online Public Access Catalogs (OPACs). In a traditional library, this problem is mitigated when the student goes to the shelf looking for any one of the books and suddenly discovers a much wider selection than the keyword search had led him to expect. But in a digital library, the issue of non-retrieval can be more problematic, as studies have indicated. Micco and Popp reported that, in a study funded partly by the U.S. Department of Education, 65 of 73 unskilled users searching for material on U.S./Soviet foreign relations found some material but never realized they had missed a large percentage of what was in the database.
  11. Gelernter, J.: Visual classification with information visualization (Infoviz) for digital library collections (2007) 0.02
    0.017707217 = product of:
      0.07968248 = sum of:
        0.05872617 = weight(_text_:applications in 423) [ClassicSimilarity], result of:
          0.05872617 = score(doc=423,freq=2.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.34048924 = fieldWeight in 423, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.0546875 = fieldNorm(doc=423)
        0.020956306 = weight(_text_:of in 423) [ClassicSimilarity], result of:
          0.020956306 = score(doc=423,freq=16.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.34207192 = fieldWeight in 423, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=423)
      0.22222222 = coord(2/9)
    
    Abstract
    The purpose of information visualization (infoviz) is to show information graphically. That purpose is often obscured by infoviz designs that are not well understood in practice. This paper offers an overview of infoviz culled from the literature on applications of information visualization for the digital library: how the clustering works that creates the topics and those topics are represented graphically. It presents a taxonomy of infoviz designs in one, two and three dimensions. It is suggested that user evaluations of infoviz designs might be used to enrich infoviz theory and, whether through application of the theory or through application of user remarks, developers might improve infoviz interface comprehensibility. Design recommendations are made in an effort to improve weaknesses and capitalize on strengths of present interfaces in representing knowledge visually.
  12. Eckert, K.; Pfeffer, M.; Stuckenschmidt, H.: Assessing thesaurus-based annotations for semantic search applications (2008) 0.02
    0.017406443 = product of:
      0.078329 = sum of:
        0.05872617 = weight(_text_:applications in 1528) [ClassicSimilarity], result of:
          0.05872617 = score(doc=1528,freq=2.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.34048924 = fieldWeight in 1528, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1528)
        0.01960283 = weight(_text_:of in 1528) [ClassicSimilarity], result of:
          0.01960283 = score(doc=1528,freq=14.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.31997898 = fieldWeight in 1528, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1528)
      0.22222222 = coord(2/9)
    
    Abstract
    Statistical methods for automated document indexing are becoming an alternative to the manual assignment of keywords. We argue that the quality of the thesaurus used as a basis for indexing in regard to its ability to adequately cover the contents to be indexed and as a basis for the specific indexing method used is of crucial importance in automatic indexing. We present an interactive tool for thesaurus evaluation that is based on a combination of statistical measures and appropriate visualisation techniques that supports the detection of potential problems in a thesaurus. We describe the methods used and show that the tool supports the detection and correction of errors, leading to a better indexing result.
    Source
    International Journal of Metadata, Semantics and Ontologies. 3(2008) no.1, S.53-67
  13. Lin, X.; Bui, Y.: Information visualization (2009) 0.02
    0.017406443 = product of:
      0.078329 = sum of:
        0.05872617 = weight(_text_:applications in 3818) [ClassicSimilarity], result of:
          0.05872617 = score(doc=3818,freq=2.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.34048924 = fieldWeight in 3818, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3818)
        0.01960283 = weight(_text_:of in 3818) [ClassicSimilarity], result of:
          0.01960283 = score(doc=3818,freq=14.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.31997898 = fieldWeight in 3818, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3818)
      0.22222222 = coord(2/9)
    
    Abstract
    The goal of information visualization (IV) is to amplify human cognition through computer-generated, interactive, and visual data representation. By combining the computational power with human perceptional and associative capabilities, IV will make it easier for users to navigate through large amounts of information, discover patterns or hidden structures of the information, and understand semantics of the information space. This entry reviews the history and background of IV and discusses its basic principles with pointers to relevant resources. The entry also summarizes major IV techniques and toolkits and shows various examples of IV applications.
    Source
    Encyclopedia of library and information sciences. 3rd ed. Ed.: M.J. Bates
  14. Yi, K.; Chan, L.M.: ¬A visualization software tool for Library of Congress Subject Headings (2008) 0.02
    0.017079117 = product of:
      0.076856025 = sum of:
        0.019052157 = weight(_text_:of in 2503) [ClassicSimilarity], result of:
          0.019052157 = score(doc=2503,freq=18.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.3109903 = fieldWeight in 2503, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=2503)
        0.05780387 = weight(_text_:software in 2503) [ClassicSimilarity], result of:
          0.05780387 = score(doc=2503,freq=4.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.3719205 = fieldWeight in 2503, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=2503)
      0.22222222 = coord(2/9)
    
    Content
    The aim of this study is to develop a software tool, VisuaLCSH, for effective searching, browsing, and maintenance of LCSH. This tool enables visualizing subject headings and hierarchical structures implied and embedded in LCSH. A conceptual framework for converting the hierarchical structure of headings in LCSH to an explicit tree structure is proposed, described, and implemented. The highlights of VisuaLCSH are summarized below: 1) revealing multiple aspects of a heading; 2) normalizing the hierarchical relationships in LCSH; 3) showing multi-level hierarchies in LCSH sub-trees; 4) improving the navigational function of LCSH in retrieval; and 5) enabling the implementation of generic search, i.e., the 'exploding' feature, in searching LCSH.
    Source
    Culture and identity in knowledge organization: Proceedings of the Tenth International ISKO Conference 5-8 August 2008, Montreal, Canada. Ed. by Clément Arsenault and Joseph T. Tennis
  15. Wu, I.-C.; Vakkari, P.: Effects of subject-oriented visualization tools on search by novices and intermediates (2018) 0.02
    0.015181154 = product of:
      0.045543462 = sum of:
        0.011833867 = weight(_text_:of in 4573) [ClassicSimilarity], result of:
          0.011833867 = score(doc=4573,freq=10.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.19316542 = fieldWeight in 4573, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4573)
        0.020439833 = weight(_text_:systems in 4573) [ClassicSimilarity], result of:
          0.020439833 = score(doc=4573,freq=2.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.1697705 = fieldWeight in 4573, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4573)
        0.013269759 = product of:
          0.026539518 = sum of:
            0.026539518 = weight(_text_:22 in 4573) [ClassicSimilarity], result of:
              0.026539518 = score(doc=4573,freq=2.0), product of:
                0.13719016 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03917671 = queryNorm
                0.19345059 = fieldWeight in 4573, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4573)
          0.5 = coord(1/2)
      0.33333334 = coord(3/9)
    
    Abstract
    This study explores how user subject knowledge influences search task processes and outcomes, as well as how search behavior is influenced by subject-oriented information visualization (IV) tools. To enable integrated searches, the proposed WikiMap + integrates search functions and IV tools (i.e., a topic network and hierarchical topic tree) and gathers information from Wikipedia pages and Google Search results. To evaluate the effectiveness of the proposed interfaces, we design subject-oriented tasks and adopt extended evaluation measures. We recruited 48 novices and 48 knowledgeable users, that is, intermediates, for the evaluation. Our results show that novices using the proposed interface demonstrate better search performance than intermediates using Wikipedia. We therefore conclude that our tools help close the gap between novices and intermediates in information searches. The results also show that intermediates can take advantage of the search tool by leveraging the IV tools to browse subtopics, and formulate better queries with less effort. We conclude that embedding the IV and the search tools in the interface can result in different search behavior but improved task performance. We provide implications to design search systems to include IV features adapted to user levels of subject knowledge to help them achieve better task performance.
    Date
    9.12.2018 16:22:25
    Source
    Journal of the Association for Information Science and Technology. 69(2018) no.12, S.1428-1445
  16. Shiri, A.; Molberg, K.: Interfaces to knowledge organization systems in Canadian digital library collections (2005) 0.01
    0.01484506 = product of:
      0.06680277 = sum of:
        0.016735615 = weight(_text_:of in 2559) [ClassicSimilarity], result of:
          0.016735615 = score(doc=2559,freq=20.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.27317715 = fieldWeight in 2559, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2559)
        0.050067157 = weight(_text_:systems in 2559) [ClassicSimilarity], result of:
          0.050067157 = score(doc=2559,freq=12.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.41585106 = fieldWeight in 2559, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2559)
      0.22222222 = coord(2/9)
    
    Abstract
    Purpose - The purpose of this paper is to report an investigation into the ways in which Canadian digital library collections have incorporated knowledge organization systems into their search interfaces. Design/methodology/approach - A combination of data-gathering techniques was used. These were as follows: a review of the literature related to the application of knowledge organization systems, deep scanning of Canadian governmental and academic institutions web sites on the web, identify and contact researchers in the area of knowledge organization, and identify and contact people in the governmental organizations who are involved in knowledge organization and information management. Findings - A total of 33 digital collections were identified that have made use of some type of knowledge organization system. Thesauri, subject heading lists and classification schemes were the widely used knowledge organization systems in the surveyed Canadian digital library collections. Research limitations/implications - The target population for this research was limited to governmental and academic digital library collections. Practical implications - An evaluation of the knowledge organization systems interfaces showed that searching, browsing and navigation facilities as well as bilingual features call for improvements. Originality/value - This research contributes to the following areas: digital libraries, knowledge organization systems and services and search interface design.
  17. Pfeffer, M.; Eckert, K.; Stuckenschmidt, H.: Visual analysis of classification systems and library collections (2008) 0.01
    0.014485389 = product of:
      0.06518425 = sum of:
        0.018934188 = weight(_text_:of in 317) [ClassicSimilarity], result of:
          0.018934188 = score(doc=317,freq=10.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.3090647 = fieldWeight in 317, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=317)
        0.046250064 = weight(_text_:systems in 317) [ClassicSimilarity], result of:
          0.046250064 = score(doc=317,freq=4.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.38414678 = fieldWeight in 317, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.0625 = fieldNorm(doc=317)
      0.22222222 = coord(2/9)
    
    Abstract
    In this demonstration we present a visual analysis approach that addresses both developers and users of hierarchical classification systems. The approach supports an intuitive understanding of the structure and current use in relation to a specific collection. We will also demonstrate its application for the development and management of library collections.
    Source
    Research and advanced technology for digital libraries : proceedings of the 12th European conference, ECDL '08, Aarhus, Denmark
  18. Choi, I.: Visualizations of cross-cultural bibliographic classification : comparative studies of the Korean Decimal Classification and the Dewey Decimal Classification (2017) 0.01
    0.013788608 = product of:
      0.062048733 = sum of:
        0.021169065 = weight(_text_:of in 3869) [ClassicSimilarity], result of:
          0.021169065 = score(doc=3869,freq=32.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.34554482 = fieldWeight in 3869, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3869)
        0.040879667 = weight(_text_:systems in 3869) [ClassicSimilarity], result of:
          0.040879667 = score(doc=3869,freq=8.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.339541 = fieldWeight in 3869, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3869)
      0.22222222 = coord(2/9)
    
    Abstract
    The changes in KO systems induced by sociocultural influences may include those in both classificatory principles and cultural features. The proposed study will examine the Korean Decimal Classification (KDC)'s adaptation of the Dewey Decimal Classification (DDC) by comparing the two systems. This case manifests the sociocultural influences on KOSs in a cross-cultural context. Therefore, the study aims at an in-depth investigation of sociocultural influences by situating a KOS in a cross-cultural environment and examining the dynamics between two classification systems designed to organize information resources in two distinct sociocultural contexts. As a preceding stage of the comparison, the analysis was conducted on the changes that result from the meeting of different sociocultural feature in a descriptive method. The analysis aims to identify variations between the two schemes in comparison of the knowledge structures of the two classifications, in terms of the quantity of class numbers that represent concepts and their relationships in each of the individual main classes. The most effective analytic strategy to show the patterns of the comparison was visualizations of similarities and differences between the two systems. Increasing or decreasing tendencies in the class through various editions were analyzed. Comparing the compositions of the main classes and distributions of concepts in the KDC and DDC discloses the differences in their knowledge structures empirically. This phase of quantitative analysis and visualizing techniques generates empirical evidence leading to interpretation.
  19. Leydesdorff, L.; Persson, O.: Mapping the geography of science : distribution patterns and networks of relations among cities and institutes (2010) 0.01
    0.013763658 = product of:
      0.06193646 = sum of:
        0.021062955 = weight(_text_:of in 3704) [ClassicSimilarity], result of:
          0.021062955 = score(doc=3704,freq=22.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.34381276 = fieldWeight in 3704, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=3704)
        0.040873505 = weight(_text_:software in 3704) [ClassicSimilarity], result of:
          0.040873505 = score(doc=3704,freq=2.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.2629875 = fieldWeight in 3704, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=3704)
      0.22222222 = coord(2/9)
    
    Abstract
    Using Google Earth, Google Maps, and/or network visualization programs such as Pajek, one can overlay the network of relations among addresses in scientific publications onto the geographic map. The authors discuss the pros and cons of various options, and provide software (freeware) for bridging existing gaps between the Science Citation Indices (Thomson Reuters) and Scopus (Elsevier), on the one hand, and these various visualization tools on the other. At the level of city names, the global map can be drawn reliably on the basis of the available address information. At the level of the names of organizations and institutes, there are problems of unification both in the ISI databases and with Scopus. Pajek enables a combination of visualization and statistical analysis, whereas the Google Maps and its derivatives provide superior tools on the Internet.
    Source
    Journal of the American Society for Information Science and Technology. 61(2010) no.8, S.1622-1634
  20. Chen, R.H.-G.; Chen, C.-M.: Visualizing the world's scientific publications (2016) 0.01
    0.013561956 = product of:
      0.0610288 = sum of:
        0.041947264 = weight(_text_:applications in 3124) [ClassicSimilarity], result of:
          0.041947264 = score(doc=3124,freq=2.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.2432066 = fieldWeight in 3124, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3124)
        0.019081537 = weight(_text_:of in 3124) [ClassicSimilarity], result of:
          0.019081537 = score(doc=3124,freq=26.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.31146988 = fieldWeight in 3124, product of:
              5.0990195 = tf(freq=26.0), with freq of:
                26.0 = termFreq=26.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3124)
      0.22222222 = coord(2/9)
    
    Abstract
    Automated methods for the analysis, modeling, and visualization of large-scale scientometric data provide measures that enable the depiction of the state of world scientific development. We aimed to integrate minimum span clustering (MSC) and minimum spanning tree methods to cluster and visualize the global pattern of scientific publications (PSP) by analyzing aggregated Science Citation Index (SCI) data from 1994 to 2011. We hypothesized that PSP clustering is mainly affected by countries' geographic location, ethnicity, and level of economic development, as indicated in previous studies. Our results showed that the 100 countries with the highest rates of publications were decomposed into 12 PSP groups and that countries within a group tended to be geographically proximal, ethnically similar, or comparable in terms of economic status. Hubs and bridging nodes in each knowledge production group were identified. The performance of each group was evaluated across 16 knowledge domains based on their specialization, volume of publications, and relative impact. Awareness of the strengths and weaknesses of each group in various knowledge domains may have useful applications for examining scientific policies, adjusting the allocation of resources, and promoting international collaboration for future developments.
    Source
    Journal of the Association for Information Science and Technology. 67(2016) no.10, S.2477-2488

Years

Languages

  • e 124
  • d 5
  • a 1
  • More… Less…

Types

  • el 22
  • b 1
  • More… Less…