Search (163 results, page 1 of 9)

  • × theme_ss:"Visualisierung"
  1. Information visualization in data mining and knowledge discovery (2002) 0.05
    0.04899536 = product of:
      0.07839258 = sum of:
        0.011807822 = weight(_text_:retrieval in 1789) [ClassicSimilarity], result of:
          0.011807822 = score(doc=1789,freq=4.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.09452859 = fieldWeight in 1789, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.015625 = fieldNorm(doc=1789)
        0.008555865 = weight(_text_:use in 1789) [ClassicSimilarity], result of:
          0.008555865 = score(doc=1789,freq=2.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.06766324 = fieldWeight in 1789, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.015625 = fieldNorm(doc=1789)
        0.019324033 = weight(_text_:of in 1789) [ClassicSimilarity], result of:
          0.019324033 = score(doc=1789,freq=150.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.2992506 = fieldWeight in 1789, product of:
              12.247449 = tf(freq=150.0), with freq of:
                150.0 = termFreq=150.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.015625 = fieldNorm(doc=1789)
        0.033110015 = sum of:
          0.004414106 = weight(_text_:on in 1789) [ClassicSimilarity], result of:
            0.004414106 = score(doc=1789,freq=2.0), product of:
              0.090823986 = queryWeight, product of:
                2.199415 = idf(docFreq=13325, maxDocs=44218)
                0.041294612 = queryNorm
              0.048600662 = fieldWeight in 1789, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                2.199415 = idf(docFreq=13325, maxDocs=44218)
                0.015625 = fieldNorm(doc=1789)
          0.02869591 = weight(_text_:line in 1789) [ClassicSimilarity], result of:
            0.02869591 = score(doc=1789,freq=2.0), product of:
              0.23157367 = queryWeight, product of:
                5.6078424 = idf(docFreq=440, maxDocs=44218)
                0.041294612 = queryNorm
              0.123916976 = fieldWeight in 1789, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.6078424 = idf(docFreq=440, maxDocs=44218)
                0.015625 = fieldNorm(doc=1789)
        0.00559485 = product of:
          0.0111897 = sum of:
            0.0111897 = weight(_text_:22 in 1789) [ClassicSimilarity], result of:
              0.0111897 = score(doc=1789,freq=2.0), product of:
                0.1446067 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.041294612 = queryNorm
                0.07738023 = fieldWeight in 1789, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.015625 = fieldNorm(doc=1789)
          0.5 = coord(1/2)
      0.625 = coord(5/8)
    
    Date
    23. 3.2008 19:10:22
    Footnote
    Rez. in: JASIST 54(2003) no.9, S.905-906 (C.A. Badurek): "Visual approaches for knowledge discovery in very large databases are a prime research need for information scientists focused an extracting meaningful information from the ever growing stores of data from a variety of domains, including business, the geosciences, and satellite and medical imagery. This work presents a summary of research efforts in the fields of data mining, knowledge discovery, and data visualization with the goal of aiding the integration of research approaches and techniques from these major fields. The editors, leading computer scientists from academia and industry, present a collection of 32 papers from contributors who are incorporating visualization and data mining techniques through academic research as well application development in industry and government agencies. Information Visualization focuses upon techniques to enhance the natural abilities of humans to visually understand data, in particular, large-scale data sets. It is primarily concerned with developing interactive graphical representations to enable users to more intuitively make sense of multidimensional data as part of the data exploration process. It includes research from computer science, psychology, human-computer interaction, statistics, and information science. Knowledge Discovery in Databases (KDD) most often refers to the process of mining databases for previously unknown patterns and trends in data. Data mining refers to the particular computational methods or algorithms used in this process. The data mining research field is most related to computational advances in database theory, artificial intelligence and machine learning. This work compiles research summaries from these main research areas in order to provide "a reference work containing the collection of thoughts and ideas of noted researchers from the fields of data mining and data visualization" (p. 8). It addresses these areas in three main sections: the first an data visualization, the second an KDD and model visualization, and the last an using visualization in the knowledge discovery process. The seven chapters of Part One focus upon methodologies and successful techniques from the field of Data Visualization. Hoffman and Grinstein (Chapter 2) give a particularly good overview of the field of data visualization and its potential application to data mining. An introduction to the terminology of data visualization, relation to perceptual and cognitive science, and discussion of the major visualization display techniques are presented. Discussion and illustration explain the usefulness and proper context of such data visualization techniques as scatter plots, 2D and 3D isosurfaces, glyphs, parallel coordinates, and radial coordinate visualizations. Remaining chapters present the need for standardization of visualization methods, discussion of user requirements in the development of tools, and examples of using information visualization in addressing research problems.
    In 13 chapters, Part Two provides an introduction to KDD, an overview of data mining techniques, and examples of the usefulness of data model visualizations. The importance of visualization throughout the KDD process is stressed in many of the chapters. In particular, the need for measures of visualization effectiveness, benchmarking for identifying best practices, and the use of standardized sample data sets is convincingly presented. Many of the important data mining approaches are discussed in this complementary context. Cluster and outlier detection, classification techniques, and rule discovery algorithms are presented as the basic techniques common to the KDD process. The potential effectiveness of using visualization in the data modeling process are illustrated in chapters focused an using visualization for helping users understand the KDD process, ask questions and form hypotheses about their data, and evaluate the accuracy and veracity of their results. The 11 chapters of Part Three provide an overview of the KDD process and successful approaches to integrating KDD, data mining, and visualization in complementary domains. Rhodes (Chapter 21) begins this section with an excellent overview of the relation between the KDD process and data mining techniques. He states that the "primary goals of data mining are to describe the existing data and to predict the behavior or characteristics of future data of the same type" (p. 281). These goals are met by data mining tasks such as classification, regression, clustering, summarization, dependency modeling, and change or deviation detection. Subsequent chapters demonstrate how visualization can aid users in the interactive process of knowledge discovery by graphically representing the results from these iterative tasks. Finally, examples of the usefulness of integrating visualization and data mining tools in the domain of business, imagery and text mining, and massive data sets are provided. This text concludes with a thorough and useful 17-page index and lengthy yet integrating 17-page summary of the academic and industrial backgrounds of the contributing authors. A 16-page set of color inserts provide a better representation of the visualizations discussed, and a URL provided suggests that readers may view all the book's figures in color on-line, although as of this submission date it only provides access to a summary of the book and its contents. The overall contribution of this work is its focus an bridging two distinct areas of research, making it a valuable addition to the Morgan Kaufmann Series in Database Management Systems. The editors of this text have met their main goal of providing the first textbook integrating knowledge discovery, data mining, and visualization. Although it contributes greatly to our under- standing of the development and current state of the field, a major weakness of this text is that there is no concluding chapter to discuss the contributions of the sum of these contributed papers or give direction to possible future areas of research. "Integration of expertise between two different disciplines is a difficult process of communication and reeducation. Integrating data mining and visualization is particularly complex because each of these fields in itself must draw an a wide range of research experience" (p. 300). Although this work contributes to the crossdisciplinary communication needed to advance visualization in KDD, a more formal call for an interdisciplinary research agenda in a concluding chapter would have provided a more satisfying conclusion to a very good introductory text.
    With contributors almost exclusively from the computer science field, the intended audience of this work is heavily slanted towards a computer science perspective. However, it is highly readable and provides introductory material that would be useful to information scientists from a variety of domains. Yet, much interesting work in information visualization from other fields could have been included giving the work more of an interdisciplinary perspective to complement their goals of integrating work in this area. Unfortunately, many of the application chapters are these, shallow, and lack complementary illustrations of visualization techniques or user interfaces used. However, they do provide insight into the many applications being developed in this rapidly expanding field. The authors have successfully put together a highly useful reference text for the data mining and information visualization communities. Those interested in a good introduction and overview of complementary research areas in these fields will be satisfied with this collection of papers. The focus upon integrating data visualization with data mining complements texts in each of these fields, such as Advances in Knowledge Discovery and Data Mining (Fayyad et al., MIT Press) and Readings in Information Visualization: Using Vision to Think (Card et. al., Morgan Kauffman). This unique work is a good starting point for future interaction between researchers in the fields of data visualization and data mining and makes a good accompaniment for a course focused an integrating these areas or to the main reference texts in these fields."
    RSWK
    Information Retrieval (BVB)
    Subject
    Information Retrieval (BVB)
  2. Zhu, Y.; Yan, E.; Song, I.-Y..: ¬The use of a graph-based system to improve bibliographic information retrieval : system design, implementation, and evaluation (2017) 0.05
    0.04729724 = product of:
      0.09459448 = sum of:
        0.050096344 = weight(_text_:retrieval in 3356) [ClassicSimilarity], result of:
          0.050096344 = score(doc=3356,freq=8.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.40105087 = fieldWeight in 3356, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=3356)
        0.025667597 = weight(_text_:use in 3356) [ClassicSimilarity], result of:
          0.025667597 = score(doc=3356,freq=2.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.20298971 = fieldWeight in 3356, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.046875 = fieldNorm(doc=3356)
        0.009466803 = weight(_text_:of in 3356) [ClassicSimilarity], result of:
          0.009466803 = score(doc=3356,freq=4.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.14660224 = fieldWeight in 3356, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=3356)
        0.009363732 = product of:
          0.018727465 = sum of:
            0.018727465 = weight(_text_:on in 3356) [ClassicSimilarity], result of:
              0.018727465 = score(doc=3356,freq=4.0), product of:
                0.090823986 = queryWeight, product of:
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.041294612 = queryNorm
                0.20619515 = fieldWeight in 3356, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3356)
          0.5 = coord(1/2)
      0.5 = coord(4/8)
    
    Abstract
    In this article, we propose a graph-based interactive bibliographic information retrieval system-GIBIR. GIBIR provides an effective way to retrieve bibliographic information. The system represents bibliographic information as networks and provides a form-based query interface. Users can develop their queries interactively by referencing the system-generated graph queries. Complex queries such as "papers on information retrieval, which were cited by John's papers that had been presented in SIGIR" can be effectively answered by the system. We evaluate the proposed system by developing another relational database-based bibliographic information retrieval system with the same interface and functions. Experiment results show that the proposed system executes the same queries much faster than the relational database-based system, and on average, our system reduced the execution time by 72% (for 3-node query), 89% (for 4-node query), and 99% (for 5-node query).
    Source
    Journal of the Association for Information Science and Technology. 68(2017) no.2, S.480-490
  3. Julien, C.-A.; Leide, J.E.; Bouthillier, F.: Controlled user evaluations of information visualization interfaces for text retrieval : literature review and meta-analysis (2008) 0.05
    0.046821814 = product of:
      0.09364363 = sum of:
        0.035423465 = weight(_text_:retrieval in 1718) [ClassicSimilarity], result of:
          0.035423465 = score(doc=1718,freq=4.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.2835858 = fieldWeight in 1718, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=1718)
        0.025667597 = weight(_text_:use in 1718) [ClassicSimilarity], result of:
          0.025667597 = score(doc=1718,freq=2.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.20298971 = fieldWeight in 1718, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.046875 = fieldNorm(doc=1718)
        0.023188837 = weight(_text_:of in 1718) [ClassicSimilarity], result of:
          0.023188837 = score(doc=1718,freq=24.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.3591007 = fieldWeight in 1718, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=1718)
        0.009363732 = product of:
          0.018727465 = sum of:
            0.018727465 = weight(_text_:on in 1718) [ClassicSimilarity], result of:
              0.018727465 = score(doc=1718,freq=4.0), product of:
                0.090823986 = queryWeight, product of:
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.041294612 = queryNorm
                0.20619515 = fieldWeight in 1718, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1718)
          0.5 = coord(1/2)
      0.5 = coord(4/8)
    
    Abstract
    This review describes experimental designs (users, search tasks, measures, etc.) used by 31 controlled user studies of information visualization (IV) tools for textual information retrieval (IR) and a meta-analysis of the reported statistical effects. Comparable experimental designs allow research designers to compare their results with other reports, and support the development of experimentally verified design guidelines concerning which IV techniques are better suited to which types of IR tasks. The studies generally use a within-subject design with 15 or more undergraduate students performing browsing to known-item tasks on sets of at least 1,000 full-text articles or Web pages on topics of general interest/news. Results of the meta-analysis (N = 8) showed no significant effects of the IV tool as compared with a text-only equivalent, but the set shows great variability suggesting an inadequate basis of comparison. Experimental design recommendations are provided which would support comparison of existing IV tools for IR usability testing.
    Source
    Journal of the American Society for Information Science and Technology. 59(2008) no.6, S.1012-1024
  4. Koshman, S.: Comparing usability between a visualization and text-based system for information retrieval (2004) 0.05
    0.045902062 = product of:
      0.091804124 = sum of:
        0.04338471 = weight(_text_:retrieval in 4424) [ClassicSimilarity], result of:
          0.04338471 = score(doc=4424,freq=6.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.34732026 = fieldWeight in 4424, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=4424)
        0.025667597 = weight(_text_:use in 4424) [ClassicSimilarity], result of:
          0.025667597 = score(doc=4424,freq=2.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.20298971 = fieldWeight in 4424, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.046875 = fieldNorm(doc=4424)
        0.013388081 = weight(_text_:of in 4424) [ClassicSimilarity], result of:
          0.013388081 = score(doc=4424,freq=8.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.20732689 = fieldWeight in 4424, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=4424)
        0.009363732 = product of:
          0.018727465 = sum of:
            0.018727465 = weight(_text_:on in 4424) [ClassicSimilarity], result of:
              0.018727465 = score(doc=4424,freq=4.0), product of:
                0.090823986 = queryWeight, product of:
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.041294612 = queryNorm
                0.20619515 = fieldWeight in 4424, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4424)
          0.5 = coord(1/2)
      0.5 = coord(4/8)
    
    Abstract
    This investigation tested the designer assumption that VIBE is a tool for an expert user and asked: what are the effects of user expertise on usability when VIBE's non-traditional interface is compared with a more traditional text-based interface? Three user groups - novices, online searching experts, and VIBE system experts - totaling 31 participants, were asked to use and compare VIBE to a more traditional text-based system, askSam. No significant differences were found; however, significant performance differences were found for some tasks on the two systems. Participants understood the basic principles underlying VIBE although they generally favored the askSam system. The findings suggest that VIBE is a learnable system and its components have pragmatic application to the development of visualized information retrieval systems. Further research is recommended to maximize the retrieval potential of IR visualization systems.
    Source
    Journal of documentation. 60(2004) no.5, S.565-580
  5. Quirin, A.; Cordón, O.; Santamaría, J.; Vargas-Quesada, B.; Moya-Anegón, F.: ¬A new variant of the Pathfinder algorithm to generate large visual science maps in cubic time (2008) 0.04
    0.044218257 = product of:
      0.11791535 = sum of:
        0.02963839 = weight(_text_:use in 2112) [ClassicSimilarity], result of:
          0.02963839 = score(doc=2112,freq=6.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.23439234 = fieldWeight in 2112, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.03125 = fieldNorm(doc=2112)
        0.018400159 = weight(_text_:of in 2112) [ClassicSimilarity], result of:
          0.018400159 = score(doc=2112,freq=34.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.28494355 = fieldWeight in 2112, product of:
              5.8309517 = tf(freq=34.0), with freq of:
                34.0 = termFreq=34.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=2112)
        0.0698768 = sum of:
          0.012484977 = weight(_text_:on in 2112) [ClassicSimilarity], result of:
            0.012484977 = score(doc=2112,freq=4.0), product of:
              0.090823986 = queryWeight, product of:
                2.199415 = idf(docFreq=13325, maxDocs=44218)
                0.041294612 = queryNorm
              0.13746344 = fieldWeight in 2112, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                2.199415 = idf(docFreq=13325, maxDocs=44218)
                0.03125 = fieldNorm(doc=2112)
          0.05739182 = weight(_text_:line in 2112) [ClassicSimilarity], result of:
            0.05739182 = score(doc=2112,freq=2.0), product of:
              0.23157367 = queryWeight, product of:
                5.6078424 = idf(docFreq=440, maxDocs=44218)
                0.041294612 = queryNorm
              0.24783395 = fieldWeight in 2112, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.6078424 = idf(docFreq=440, maxDocs=44218)
                0.03125 = fieldNorm(doc=2112)
      0.375 = coord(3/8)
    
    Abstract
    In the last few years, there is an increasing interest to generate visual representations of very large scientific domains. A methodology based on the combined use of ISI-JCR category cocitation and social networks analysis through the use of the Pathfinder algorithm has demonstrated its ability to achieve high quality, schematic visualizations for these kinds of domains. Now, the next step would be to generate these scientograms in an on-line fashion. To do so, there is a need to significantly decrease the run time of the latter pruning technique when working with category cocitation matrices of a large dimension like the ones handled in these large domains (Pathfinder has a time complexity order of O(n4), with n being the number of categories in the cocitation matrix, i.e., the number of nodes in the network). Although a previous improvement called Binary Pathfinder has already been proposed to speed up the original algorithm, its significant time complexity reduction is not enough for that aim. In this paper, we make use of a different shortest path computation from classical approaches in computer science graph theory to propose a new variant of the Pathfinder algorithm which allows us to reduce its time complexity in one order of magnitude, O(n3), and thus to significantly decrease the run time of the implementation when applied to large scientific domains considering the parameter q = n - 1. Besides, the new algorithm has a much simpler structure than the Binary Pathfinder as well as it saves a significant amount of memory with respect to the original Pathfinder by reducing the space complexity to the need of just storing two matrices. An experimental comparison will be developed using large networks from real-world domains to show the good performance of the new proposal.
  6. Golub, K.; Ziolkowski, P.M.; Zlodi, G.: Organizing subject access to cultural heritage in Swedish online museums (2022) 0.04
    0.042145163 = product of:
      0.084290326 = sum of:
        0.033397563 = weight(_text_:retrieval in 688) [ClassicSimilarity], result of:
          0.033397563 = score(doc=688,freq=8.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.26736724 = fieldWeight in 688, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=688)
        0.024199642 = weight(_text_:use in 688) [ClassicSimilarity], result of:
          0.024199642 = score(doc=688,freq=4.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.19138055 = fieldWeight in 688, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.03125 = fieldNorm(doc=688)
        0.020450631 = weight(_text_:of in 688) [ClassicSimilarity], result of:
          0.020450631 = score(doc=688,freq=42.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.31669703 = fieldWeight in 688, product of:
              6.4807405 = tf(freq=42.0), with freq of:
                42.0 = termFreq=42.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=688)
        0.0062424885 = product of:
          0.012484977 = sum of:
            0.012484977 = weight(_text_:on in 688) [ClassicSimilarity], result of:
              0.012484977 = score(doc=688,freq=4.0), product of:
                0.090823986 = queryWeight, product of:
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.041294612 = queryNorm
                0.13746344 = fieldWeight in 688, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.03125 = fieldNorm(doc=688)
          0.5 = coord(1/2)
      0.5 = coord(4/8)
    
    Abstract
    Purpose The study aims to paint a representative picture of the current state of search interfaces of Swedish online museum collections, focussing on search functionalities with particular reference to subject searching, as well as the use of controlled vocabularies, with the purpose of identifying which improvements of the search interfaces are needed to ensure high-quality information retrieval for the end user. Design/methodology/approach In the first step, a set of 21 search interface criteria was identified, based on related research and current standards in the domain of cultural heritage knowledge organization. Secondly, a complete set of Swedish museums that provide online access to their collections was identified, comprising nine cross-search services and 91 individual museums' websites. These 100 websites were each evaluated against the 21 criteria, between 1 July and 31 August 2020. Findings Although many standards and guidelines are in place to ensure quality-controlled subject indexing, which in turn support information retrieval of relevant resources (as individual or full search results), the study shows that they are not broadly implemented, resulting in information retrieval failures for the end user. The study also demonstrates a strong need for the implementation of controlled vocabularies in these museums. Originality/value This study is a rare piece of research which examines subject searching in online museums; the 21 search criteria and their use in the analysis of the complete set of online collections of a country represents a considerable and unique contribution to the fields of knowledge organization and information retrieval of cultural heritage. Its particular value lies in showing how the needs of end users, many of which are documented and reflected in international standards and guidelines, should be taken into account in designing search tools for these museums; especially so in subject searching, which is the most complex and yet the most common type of search. Much effort has been invested into digitizing cultural heritage collections, but access to them is hindered by poor search functionality. This study identifies which are the most important aspects to improve.
    Source
    Journal of documentation. 78(2022) no.7, S.211-247
  7. Eckert, K.: Thesaurus analysis and visualization in semantic search applications (2007) 0.04
    0.041035485 = product of:
      0.08207097 = sum of:
        0.029519552 = weight(_text_:retrieval in 3222) [ClassicSimilarity], result of:
          0.029519552 = score(doc=3222,freq=4.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.23632148 = fieldWeight in 3222, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3222)
        0.021389665 = weight(_text_:use in 3222) [ClassicSimilarity], result of:
          0.021389665 = score(doc=3222,freq=2.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.1691581 = fieldWeight in 3222, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3222)
        0.021604925 = weight(_text_:of in 3222) [ClassicSimilarity], result of:
          0.021604925 = score(doc=3222,freq=30.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.33457235 = fieldWeight in 3222, product of:
              5.477226 = tf(freq=30.0), with freq of:
                30.0 = termFreq=30.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3222)
        0.00955682 = product of:
          0.01911364 = sum of:
            0.01911364 = weight(_text_:on in 3222) [ClassicSimilarity], result of:
              0.01911364 = score(doc=3222,freq=6.0), product of:
                0.090823986 = queryWeight, product of:
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.041294612 = queryNorm
                0.21044704 = fieldWeight in 3222, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3222)
          0.5 = coord(1/2)
      0.5 = coord(4/8)
    
    Abstract
    The use of thesaurus-based indexing is a common approach for increasing the performance of information retrieval. In this thesis, we examine the suitability of a thesaurus for a given set of information and evaluate improvements of existing thesauri to get better search results. On this area, we focus on two aspects: 1. We demonstrate an analysis of the indexing results achieved by an automatic document indexer and the involved thesaurus. 2. We propose a method for thesaurus evaluation which is based on a combination of statistical measures and appropriate visualization techniques that support the detection of potential problems in a thesaurus. In this chapter, we give an overview of the context of our work. Next, we briefly outline the basics of thesaurus-based information retrieval and describe the Collexis Engine that was used for our experiments. In Chapter 3, we describe two experiments in automatically indexing documents in the areas of medicine and economics with corresponding thesauri and compare the results to available manual annotations. Chapter 4 describes methods for assessing thesauri and visualizing the result in terms of a treemap. We depict examples of interesting observations supported by the method and show that we actually find critical problems. We conclude with a discussion of open questions and future research in Chapter 5.
  8. Enser, P.: ¬The evolution of visual information retrieval (2009) 0.04
    0.040200733 = product of:
      0.10720195 = sum of:
        0.07158111 = weight(_text_:retrieval in 3659) [ClassicSimilarity], result of:
          0.07158111 = score(doc=3659,freq=12.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.5730491 = fieldWeight in 3659, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3659)
        0.024696484 = weight(_text_:of in 3659) [ClassicSimilarity], result of:
          0.024696484 = score(doc=3659,freq=20.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.38244802 = fieldWeight in 3659, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3659)
        0.010924355 = product of:
          0.02184871 = sum of:
            0.02184871 = weight(_text_:on in 3659) [ClassicSimilarity], result of:
              0.02184871 = score(doc=3659,freq=4.0), product of:
                0.090823986 = queryWeight, product of:
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.041294612 = queryNorm
                0.24056101 = fieldWeight in 3659, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3659)
          0.5 = coord(1/2)
      0.375 = coord(3/8)
    
    Abstract
    This paper seeks to provide a brief overview of those developments which have taken the theory and practice of image and video retrieval into the digital age. Drawing on a voluminous literature, the context in which visual information retrieval takes place is followed by a consideration of the conceptual and practical challenges posed by the representation and recovery of visual material on the basis of its semantic content. An historical account of research endeavours in content-based retrieval, directed towards the automation of these operations in digital image scenarios, provides the main thrust of the paper. Finally, a look forwards locates visual information retrieval research within the wider context of content-based multimedia retrieval.
  9. Heo, M.; Hirtle, S.C.: ¬An empirical comparison of visualization tools to assist information retrieval on the Web (2001) 0.04
    0.039154742 = product of:
      0.078309484 = sum of:
        0.016698781 = weight(_text_:retrieval in 5215) [ClassicSimilarity], result of:
          0.016698781 = score(doc=5215,freq=2.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.13368362 = fieldWeight in 5215, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=5215)
        0.038262997 = weight(_text_:use in 5215) [ClassicSimilarity], result of:
          0.038262997 = score(doc=5215,freq=10.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.30259922 = fieldWeight in 5215, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.03125 = fieldNorm(doc=5215)
        0.018933605 = weight(_text_:of in 5215) [ClassicSimilarity], result of:
          0.018933605 = score(doc=5215,freq=36.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.2932045 = fieldWeight in 5215, product of:
              6.0 = tf(freq=36.0), with freq of:
                36.0 = termFreq=36.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=5215)
        0.004414106 = product of:
          0.008828212 = sum of:
            0.008828212 = weight(_text_:on in 5215) [ClassicSimilarity], result of:
              0.008828212 = score(doc=5215,freq=2.0), product of:
                0.090823986 = queryWeight, product of:
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.041294612 = queryNorm
                0.097201325 = fieldWeight in 5215, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5215)
          0.5 = coord(1/2)
      0.5 = coord(4/8)
    
    Abstract
    The reader of a hypertext document in a web environment, if maximum use of the document is to be obtained, must visualize the overall structure of the paths through the document as well as the document space. Graphic visualization displays of this space, produced to assist in navigation, are classified into four groups, and Heo and Hirtle compare three of these classes as to their effectiveness. Distortion displays expand regions of interest while relatively diminishing the detail of the remaining regions. This technique will show both local detail and global structure. Zoom techniques use a series of increasingly focused displays of smaller and smaller areas, and can reduce cogitative overload, but do not provide an easy movement to other parts of the total space. Expanding outline displays use a tree structure to allow movement through a hierarchy of documents, but if the organization has a wide horizontal structure, or is not particularly hierarchical in nature such display can break down. Three dimensional layouts, which are not evaluated here, place objects by location in three space, providing more information and freedom. However, the space must be represented in two dimensions resulting in difficulty in visually judging depth, size and positioning. Ten students were assigned to each of eight groups composed of viewers of the three techniques and an unassisted control group using either a large (583 selected pages) or a small (50 selected pages) web space. Sets of 10 questions, which were designed to elicit the use of a visualization tool, were provided for each space. Accuracy and time spent were extracted from a log file. Users views were also surveyed after completion. ANOVA shows significant differences in accuracy and time based upon the visualization tool in use. A Tukey test shows zoom accuracy to be significantly less than expanding outline and zoom time to be significantly greater than both the outline and control groups. Size significantly affected accuracy and time, but had no interaction with tool type. While the expanding tool class out performed zoom and distortion, its performance was not significantly different from the control group.
    Source
    Journal of the American Society for Information Science and technology. 52(2001) no.8, S.666-675
  10. Fowler, R.H.; Wilson, B.A.; Fowler, W.A.L.: Information navigator : an information system using associative networks for display and retrieval (1992) 0.04
    0.037728988 = product of:
      0.10061063 = sum of:
        0.05600942 = weight(_text_:retrieval in 919) [ClassicSimilarity], result of:
          0.05600942 = score(doc=919,freq=10.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.44838852 = fieldWeight in 919, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=919)
        0.025667597 = weight(_text_:use in 919) [ClassicSimilarity], result of:
          0.025667597 = score(doc=919,freq=2.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.20298971 = fieldWeight in 919, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.046875 = fieldNorm(doc=919)
        0.018933605 = weight(_text_:of in 919) [ClassicSimilarity], result of:
          0.018933605 = score(doc=919,freq=16.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.2932045 = fieldWeight in 919, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=919)
      0.375 = coord(3/8)
    
    Abstract
    Document retrieval is a highly interactive process dealing with large amounts of information. Visual representations can provide both a means for managing the complexity of large information structures and an interface style well suited to interactive manipulation. The system we have designed utilizes visually displayed graphic structures and a direct manipulation interface style to supply an integrated environment for retrieval. A common visually displayed network structure is used for query, document content, and term relations. A query can be modified through direct manipulation of its visual form by incorporating terms from any other information structure the system displays. An associative thesaurus of terms and an inter-document network provide information about a document collection that can complement other retrieval aids. Visualization of these large data structures makes use of fisheye views and overview diagrams to help overcome some of the inherent difficulties of orientation and navigation in large information structures.
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
  11. Zhang, J.; Mostafa, J.; Tripathy, H.: Information retrieval by semantic analysis and visualization of the concept space of D-Lib® magazine (2002) 0.04
    0.036912464 = product of:
      0.07382493 = sum of:
        0.029519552 = weight(_text_:retrieval in 1211) [ClassicSimilarity], result of:
          0.029519552 = score(doc=1211,freq=16.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.23632148 = fieldWeight in 1211, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.01953125 = fieldNorm(doc=1211)
        0.023914373 = weight(_text_:use in 1211) [ClassicSimilarity], result of:
          0.023914373 = score(doc=1211,freq=10.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.18912451 = fieldWeight in 1211, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.01953125 = fieldNorm(doc=1211)
        0.014222101 = weight(_text_:of in 1211) [ClassicSimilarity], result of:
          0.014222101 = score(doc=1211,freq=52.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.22024246 = fieldWeight in 1211, product of:
              7.2111025 = tf(freq=52.0), with freq of:
                52.0 = termFreq=52.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.01953125 = fieldNorm(doc=1211)
        0.006168901 = product of:
          0.012337802 = sum of:
            0.012337802 = weight(_text_:on in 1211) [ClassicSimilarity], result of:
              0.012337802 = score(doc=1211,freq=10.0), product of:
                0.090823986 = queryWeight, product of:
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.041294612 = queryNorm
                0.135843 = fieldWeight in 1211, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=1211)
          0.5 = coord(1/2)
      0.5 = coord(4/8)
    
    Abstract
    In this article we present a method for retrieving documents from a digital library through a visual interface based on automatically generated concepts. We used a vocabulary generation algorithm to generate a set of concepts for the digital library and a technique called the max-min distance technique to cluster them. Additionally, the concepts were visualized in a spring embedding graph layout to depict the semantic relationship among them. The resulting graph layout serves as an aid to users for retrieving documents. An online archive containing the contents of D-Lib Magazine from July 1995 to May 2002 was used to test the utility of an implemented retrieval and visualization system. We believe that the method developed and tested can be applied to many different domains to help users get a better understanding of online document collections and to minimize users' cognitive load during execution of search tasks. Over the past few years, the volume of information available through the World Wide Web has been expanding exponentially. Never has so much information been so readily available and shared among so many people. Unfortunately, the unstructured nature and huge volume of information accessible over networks have made it hard for users to sift through and find relevant information. To deal with this problem, information retrieval (IR) techniques have gained more intensive attention from both industrial and academic researchers. Numerous IR techniques have been developed to help deal with the information overload problem. These techniques concentrate on mathematical models and algorithms for retrieval. Popular IR models such as the Boolean model, the vector-space model, the probabilistic model and their variants are well established.
    From the user's perspective, however, it is still difficult to use current information retrieval systems. Users frequently have problems expressing their information needs and translating those needs into queries. This is partly due to the fact that information needs cannot be expressed appropriately in systems terms. It is not unusual for users to input search terms that are different from the index terms information systems use. Various methods have been proposed to help users choose search terms and articulate queries. One widely used approach is to incorporate into the information system a thesaurus-like component that represents both the important concepts in a particular subject area and the semantic relationships among those concepts. Unfortunately, the development and use of thesauri is not without its own problems. The thesaurus employed in a specific information system has often been developed for a general subject area and needs significant enhancement to be tailored to the information system where it is to be used. This thesaurus development process, if done manually, is both time consuming and labor intensive. Usage of a thesaurus in searching is complex and may raise barriers for the user. For illustration purposes, let us consider two scenarios of thesaurus usage. In the first scenario the user inputs a search term and the thesaurus then displays a matching set of related terms. Without an overview of the thesaurus - and without the ability to see the matching terms in the context of other terms - it may be difficult to assess the quality of the related terms in order to select the correct term. In the second scenario the user browses the whole thesaurus, which is organized as in an alphabetically ordered list. The problem with this approach is that the list may be long, and neither does it show users the global semantic relationship among all the listed terms.
    Nevertheless, because thesaurus use has shown to improve retrieval, for our method we integrate functions in the search interface that permit users to explore built-in search vocabularies to improve retrieval from digital libraries. Our method automatically generates the terms and their semantic relationships representing relevant topics covered in a digital library. We call these generated terms the "concepts", and the generated terms and their semantic relationships we call the "concept space". Additionally, we used a visualization technique to display the concept space and allow users to interact with this space. The automatically generated term set is considered to be more representative of subject area in a corpus than an "externally" imposed thesaurus, and our method has the potential of saving a significant amount of time and labor for those who have been manually creating thesauri as well. Information visualization is an emerging discipline and developed very quickly in the last decade. With growing volumes of documents and associated complexities, information visualization has become increasingly important. Researchers have found information visualization to be an effective way to use and understand information while minimizing a user's cognitive load. Our work was based on an algorithmic approach of concept discovery and association. Concepts are discovered using an algorithm based on an automated thesaurus generation procedure. Subsequently, similarities among terms are computed using the cosine measure, and the associations among terms are established using a method known as max-min distance clustering. The concept space is then visualized in a spring embedding graph, which roughly shows the semantic relationships among concepts in a 2-D visual representation. The semantic space of the visualization is used as a medium for users to retrieve the desired documents. In the remainder of this article, we present our algorithmic approach of concept generation and clustering, followed by description of the visualization technique and interactive interface. The paper ends with key conclusions and discussions on future work.
    Content
    The JAVA applet is available at <http://ella.slis.indiana.edu/~junzhang/dlib/IV.html>. A prototype of this interface has been developed and is available at <http://ella.slis.indiana.edu/~junzhang/dlib/IV.html>. The D-Lib search interface is available at <http://www.dlib.org/Architext/AT-dlib2query.html>.
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
  12. Koch, T.; Golub, K.; Ardö, A.: Users browsing behaviour in a DDC-based Web service : a log analysis (2006) 0.04
    0.036866955 = product of:
      0.07373391 = sum of:
        0.025048172 = weight(_text_:retrieval in 2234) [ClassicSimilarity], result of:
          0.025048172 = score(doc=2234,freq=2.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.20052543 = fieldWeight in 2234, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=2234)
        0.025667597 = weight(_text_:use in 2234) [ClassicSimilarity], result of:
          0.025667597 = score(doc=2234,freq=2.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.20298971 = fieldWeight in 2234, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.046875 = fieldNorm(doc=2234)
        0.016396983 = weight(_text_:of in 2234) [ClassicSimilarity], result of:
          0.016396983 = score(doc=2234,freq=12.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.25392252 = fieldWeight in 2234, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=2234)
        0.006621159 = product of:
          0.013242318 = sum of:
            0.013242318 = weight(_text_:on in 2234) [ClassicSimilarity], result of:
              0.013242318 = score(doc=2234,freq=2.0), product of:
                0.090823986 = queryWeight, product of:
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.041294612 = queryNorm
                0.14580199 = fieldWeight in 2234, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2234)
          0.5 = coord(1/2)
      0.5 = coord(4/8)
    
    Abstract
    This study explores the navigation behaviour of all users of a large web service, Renardus, using web log analysis. Renardus provides integrated searching and browsing access to quality-controlled web resources from major individual subject gateway services. The main navigation feature is subject browsing through the Dewey Decimal Classification (DDC) based on mapping of classes of resources from the distributed gateways to the DDC structure. Among the more surprising results are the hugely dominant share of browsing activities, the good use of browsing support features like the graphical fish-eye overviews, rather long and varied navigation sequences, as well as extensive hierarchical directory-style browsing through the large DDC system.
    Theme
    Klassifikationssysteme im Online-Retrieval
  13. Chowdhury, S.; Chowdhury, G.G.: Using DDC to create a visual knowledge map as an aid to online information retrieval (2004) 0.04
    0.035697732 = product of:
      0.09519395 = sum of:
        0.05009634 = weight(_text_:retrieval in 2643) [ClassicSimilarity], result of:
          0.05009634 = score(doc=2643,freq=18.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.40105084 = fieldWeight in 2643, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=2643)
        0.02963839 = weight(_text_:use in 2643) [ClassicSimilarity], result of:
          0.02963839 = score(doc=2643,freq=6.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.23439234 = fieldWeight in 2643, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.03125 = fieldNorm(doc=2643)
        0.0154592255 = weight(_text_:of in 2643) [ClassicSimilarity], result of:
          0.0154592255 = score(doc=2643,freq=24.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.23940048 = fieldWeight in 2643, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=2643)
      0.375 = coord(3/8)
    
    Abstract
    Selection of search terms in an online search environment can be facilitated by the visual display of a knowledge map showing the various concepts and their links. This paper reports an a preliminary research aimed at designing a prototype knowledge map using DDC and its visual display. The prototype knowledge map created using the Protégé and TGViz freeware has been demonstrated, and further areas of research in this field are discussed.
    Content
    1. Introduction Web search engines and digital libraries usually expect the users to use search terms that most accurately represent their information needs. Finding the most appropriate search terms to represent an information need is an age old problem in information retrieval. Keyword or phrase search may produce good search results as long as the search terms or phrase(s) match those used by the authors and have been chosen for indexing by the concerned information retrieval system. Since this does not always happen, a large number of false drops are produced by information retrieval systems. The retrieval results become worse in very large systems that deal with millions of records, such as the Web search engines and digital libraries. Vocabulary control tools are used to improve the performance of text retrieval systems. Thesauri, the most common type of vocabulary control tool used in information retrieval, appeared in the late fifties, designed for use with the emerging post-coordinate indexing systems of that time. They are used to exert terminology control in indexing, and to aid in searching by allowing the searcher to select appropriate search terms. A large volume of literature exists describing the design features, and experiments with the use, of thesauri in various types of information retrieval systems (see for example, Furnas et.al., 1987; Bates, 1986, 1998; Milstead, 1997, and Shiri et al., 2002).
    Source
    Knowledge organization and the global information society: Proceedings of the 8th International ISKO Conference 13-16 July 2004, London, UK. Ed.: I.C. McIlwaine
    Theme
    Klassifikationssysteme im Online-Retrieval
  14. Catarci, T.; Spaccapietra, S.: Visual information querying (2002) 0.03
    0.03374344 = product of:
      0.06748688 = sum of:
        0.017711733 = weight(_text_:retrieval in 4268) [ClassicSimilarity], result of:
          0.017711733 = score(doc=4268,freq=4.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.1417929 = fieldWeight in 4268, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0234375 = fieldNorm(doc=4268)
        0.012833798 = weight(_text_:use in 4268) [ClassicSimilarity], result of:
          0.012833798 = score(doc=4268,freq=2.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.101494856 = fieldWeight in 4268, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0234375 = fieldNorm(doc=4268)
        0.018024256 = weight(_text_:of in 4268) [ClassicSimilarity], result of:
          0.018024256 = score(doc=4268,freq=58.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.27912235 = fieldWeight in 4268, product of:
              7.615773 = tf(freq=58.0), with freq of:
                58.0 = termFreq=58.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0234375 = fieldNorm(doc=4268)
        0.018917091 = product of:
          0.037834182 = sum of:
            0.037834182 = weight(_text_:computers in 4268) [ClassicSimilarity], result of:
              0.037834182 = score(doc=4268,freq=2.0), product of:
                0.21710795 = queryWeight, product of:
                  5.257537 = idf(docFreq=625, maxDocs=44218)
                  0.041294612 = queryNorm
                0.17426437 = fieldWeight in 4268, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.257537 = idf(docFreq=625, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=4268)
          0.5 = coord(1/2)
      0.5 = coord(4/8)
    
    Abstract
    Computers have become our companions in many of the activities we pursue in our life. They assist us, in particular, in searching relevant information that is needed to perform a variety of tasks, from professional usage to personal entertainment. They hold this information in a huge number of heterogeneous sources, either dedicated to a specific user community (e.g., enterprise databases) or maintained for the general public (e.g., websites and digital libraries). Whereas progress in basic information technology is nowadays capable of guaranteeing effective information management, information retrieval and dissemination has become a core issue that needs further accomplishments to achieve user satisfaction. The research communities in databases, information retrieval, information visualization, and human-computer interaction have already largely investigated these domains. However, the technical environment has so dramatically evolved in recent years, inducing a parallel and very significant evolution in user habits and expectations, that new approaches are definitely needed to meet current demand. One of the most evident and significant changes is the human-computer interaction paradigm. Traditional interactions relayed an programming to express user information requirements in formal code and an textual output to convey to users the information extracted by the system. Except for professional data-intensive application frameworks, still in the hands of computer speciahsts, we have basically moved away from this pattern both in terms of expressing information requests and conveying results. The new goal is direct interaction with the final user (the person who is looking for information and is not necessarily familiar with computer technology). The key motto to achieve this is "go visual." The well-known high bandwidth of the human-vision channel allows both recognition and understanding of large quantities of information in no more than a few seconds. Thus, for instance, if the result of an information request can be organized as a visual display, or a sequence of visual displays, the information throughput is immensely superior to the one that can be achieved using textual support. User interaction becomes an iterative query-answer game that very rapidly leads to the desired final result. Conversely, the system can provide efficient visual support for easy query formulation. Displaying a visual representation of the information space, for instance, lets users directly point at the information they are looking for, without any need to be trained into the complex syntax of current query languages. Alternatively, users can navigate in the information space, following visible paths that will lead them to the targeted items. Again, thanks to the visual support, users are able to easily understand how to formulate queries and they are likely to achieve the task more rapidly and less prone to errors than with traditional textual interaction modes.
    The two facets of "going visual" are usually referred to as visual query systems, for query formulation, and information visualization, for result display. Visual Query Systems (VQSs) are defined as systems for querying databases that use a visual representation to depict the domain of interest and express related requests. VQSs provide both a language to express the queries in a visual format and a variety of functionalities to facilitate user-system interaction. As such, they are oriented toward a wide spectrum of users, especially novices who have limited computer expertise and generally ignore the inner structure of the accessed database. Information visualization, an increasingly important subdiscipline within the field of Human-Computer Interaction (HCI), focuses an visual mechanisms designed to communicate clearly to the user the structure of information and improve an the cost of accessing large data repositories. In printed form, information visualization has included the display of numerical data (e.g., bar charts, plot charts, pie charts), combinatorial relations (e.g., drawings of graphs), and geographic data (e.g., encoded maps). In addition to these "static" displays, computer-based systems, such as the Information Visualizer and Dynamic Queries, have coupled powerful visualization techniques (e.g., 3D, animation) with near real-time interactivity (i.e., the ability of the system to respond quickly to the user's direct manipulation commands). Information visualization is tightly combined with querying capabilities in some recent database-centered approaches. More opportunities for information visualization in a database environment may be found today in data mining and data warehousing applications, which typically access large data repositories. The enormous quantity of information sources an the World-Wide Web (WWW) available to users with diverse capabilities also calls for visualization techniques. In this article, we survey the main features and main proposals for visual query systems and touch upon the visualization of results mainly discussing traditional visualization forms. A discussion of modern database visualization techniques may be found elsewhere. Many related articles by Daniel Keim are available at http://www. informatik.uni-halle.de/dbs/publications.html.
    Source
    Encyclopedia of library and information science. Vol.72, [=Suppl.35]
  15. Dushay, N.: Visualizing bibliographic metadata : a virtual (book) spine viewer (2004) 0.03
    0.03333898 = product of:
      0.06667796 = sum of:
        0.012833798 = weight(_text_:use in 1197) [ClassicSimilarity], result of:
          0.012833798 = score(doc=1197,freq=2.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.101494856 = fieldWeight in 1197, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1197)
        0.018332386 = weight(_text_:of in 1197) [ClassicSimilarity], result of:
          0.018332386 = score(doc=1197,freq=60.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.28389403 = fieldWeight in 1197, product of:
              7.745967 = tf(freq=60.0), with freq of:
                60.0 = termFreq=60.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1197)
        0.008758971 = product of:
          0.017517941 = sum of:
            0.017517941 = weight(_text_:on in 1197) [ClassicSimilarity], result of:
              0.017517941 = score(doc=1197,freq=14.0), product of:
                0.090823986 = queryWeight, product of:
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.041294612 = queryNorm
                0.19287792 = fieldWeight in 1197, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1197)
          0.5 = coord(1/2)
        0.026752805 = product of:
          0.05350561 = sum of:
            0.05350561 = weight(_text_:computers in 1197) [ClassicSimilarity], result of:
              0.05350561 = score(doc=1197,freq=4.0), product of:
                0.21710795 = queryWeight, product of:
                  5.257537 = idf(docFreq=625, maxDocs=44218)
                  0.041294612 = queryNorm
                0.24644704 = fieldWeight in 1197, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.257537 = idf(docFreq=625, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1197)
          0.5 = coord(1/2)
      0.5 = coord(4/8)
    
    Abstract
    User interfaces for digital information discovery often require users to click around and read a lot of text in order to find the text they want to read-a process that is often frustrating and tedious. This is exacerbated because of the limited amount of text that can be displayed on a computer screen. To improve the user experience of computer mediated information discovery, information visualization techniques are applied to the digital library context, while retaining traditional information organization concepts. In this article, the "virtual (book) spine" and the virtual spine viewer are introduced. The virtual spine viewer is an application which allows users to visually explore large information spaces or collections while also allowing users to hone in on individual resources of interest. The virtual spine viewer introduced here is an alpha prototype, presented to promote discussion and further work. Information discovery changed radically with the introduction of computerized library access catalogs, the World Wide Web and its search engines, and online bookstores. Yet few instances of these technologies provide a user experience analogous to walking among well-organized, well-stocked bookshelves-which many people find useful as well as pleasurable. To put it another way, many of us have heard or voiced complaints about the paucity of "online browsing"-but what does this really mean? In traditional information spaces such as libraries, often we can move freely among the books and other resources. When we walk among organized, labeled bookshelves, we get a sense of the information space-we take in clues, perhaps unconsciously, as to the scope of the collection, the currency of resources, the frequency of their use, etc. We also enjoy unexpected discoveries such as finding an interesting resource because library staff deliberately located it near similar resources, or because it was miss-shelved, or because we saw it on a bookshelf on the way to the water fountain.
    When our experience of information discovery is mediated by a computer, we neither move ourselves nor the monitor. We have only the computer's monitor to view, and the keyboard and/or mouse to manipulate what is displayed there. Computer interfaces often reduce our ability to get a sense of the contents of a library: we don't perceive the scope of the library: its breadth, (the quantity of materials/information), its density (how full the shelves are, how thorough the collection is for individual topics), or the general audience for the materials (e.g., whether the materials are appropriate for middle school students, college professors, etc.). Additionally, many computer interfaces for information discovery require users to scroll through long lists, to click numerous navigational links and to read a lot of text to find the exact text they want to read. Text features of resources are almost always presented alphabetically, and the number of items in these alphabetical lists sometimes can be very long. Alphabetical ordering is certainly an improvement over no ordering, but it generally has no bearing on features with an inherent non-alphabetical ordering (e.g., dates of historical events), nor does it necessarily group similar items together. Alphabetical ordering of resources is analogous to one of the most familiar complaints about dictionaries: sometimes you need to know how to spell a word in order to look up its correct spelling in the dictionary. Some have used technology to replicate the appearance of physical libraries, presenting rooms of bookcases and shelves of book spines in virtual 3D environments. This approach presents a problem, as few book spines can be displayed legibly on a monitor screen. This article examines the role of book spines, call numbers, and other traditional organizational and information discovery concepts, and integrates this knowledge with information visualization techniques to show how computers and monitors can meet or exceed similar information discovery methods. The goal is to tap the unique potentials of current information visualization approaches in order to improve information discovery, offer new services, and most important of all, improve user satisfaction. We need to capitalize on what computers do well while bearing in mind their limitations. The intent is to design GUIs to optimize utility and provide a positive experience for the user.
  16. Tang, M.-C.: Browsing and searching in a faceted information space : a naturalistic study of PubMed users' interaction with a display tool (2007) 0.03
    0.03328459 = product of:
      0.08875891 = sum of:
        0.036153924 = weight(_text_:retrieval in 617) [ClassicSimilarity], result of:
          0.036153924 = score(doc=617,freq=6.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.28943354 = fieldWeight in 617, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=617)
        0.0167351 = weight(_text_:of in 617) [ClassicSimilarity], result of:
          0.0167351 = score(doc=617,freq=18.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.25915858 = fieldWeight in 617, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=617)
        0.035869885 = product of:
          0.07173977 = sum of:
            0.07173977 = weight(_text_:line in 617) [ClassicSimilarity], result of:
              0.07173977 = score(doc=617,freq=2.0), product of:
                0.23157367 = queryWeight, product of:
                  5.6078424 = idf(docFreq=440, maxDocs=44218)
                  0.041294612 = queryNorm
                0.30979243 = fieldWeight in 617, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.6078424 = idf(docFreq=440, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=617)
          0.5 = coord(1/2)
      0.375 = coord(3/8)
    
    Abstract
    The study adopts a naturalistic approach to investigate users' interaction with a browsable MeSH (medical subject headings) display designed to facilitate query construction for the PubMed bibliographic database. The purpose of the study is twofold: first, to test the usefulness of a browsable interface utilizing the principle of faceted classification; and second, to investigate users' preferred query submission methods in different problematic situations. An interface that incorporated multiple query submission methods - the conventional single-line query box as well as methods associated the faceted classification display was constructed. Participants' interactions with the interface were monitored remotely over a period of 10 weeks; information about their problematic situations and information retrieval behaviors were also collected during this time. The traditional controlled experiment was not adequate in answering the author's research questions; hence, the author provides his rationale for a naturalistic approach. The study's findings show that there is indeed a selective compatibility between query submission methods provided by the MeSH display and users' problematic situations. The query submission methods associated with the display were found to be the preferred search tools when users' information needs were vague and the search topics unfamiliar. The findings support the theoretical proposition that users engaging in an information retrieval process with a variety of problematic situations need different approaches. The author argues that rather than treat the information retrieval system as a general purpose tool, more attention should be given to the interaction between the functionality of the tool and the characteristics of users' problematic situations.
    Source
    Journal of the American Society for Information Science and Technology. 58(2007) no.13, S.1998-2006
  17. Hemmje, M.: LyberWorld - a 3D graphical user interface for fulltext retrieval (1995) 0.03
    0.031362552 = product of:
      0.083633475 = sum of:
        0.058445733 = weight(_text_:retrieval in 2385) [ClassicSimilarity], result of:
          0.058445733 = score(doc=2385,freq=8.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.46789268 = fieldWeight in 2385, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2385)
        0.017463053 = weight(_text_:of in 2385) [ClassicSimilarity], result of:
          0.017463053 = score(doc=2385,freq=10.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.2704316 = fieldWeight in 2385, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2385)
        0.007724685 = product of:
          0.01544937 = sum of:
            0.01544937 = weight(_text_:on in 2385) [ClassicSimilarity], result of:
              0.01544937 = score(doc=2385,freq=2.0), product of:
                0.090823986 = queryWeight, product of:
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.041294612 = queryNorm
                0.17010231 = fieldWeight in 2385, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2385)
          0.5 = coord(1/2)
      0.375 = coord(3/8)
    
    Abstract
    LyberWorld is a prototype IR user interface. It implements visualizations of an abstract information space: fulltext. The video demonstrates a visual user interface for the probabilistic fulltext retrieval system INQUERY. Visualizations are used to communicate information search and browsing activities in a natural way by applying metaphors of spatial navigation in abstract information spaces. Visualization tools for exploring information spaces and judging relevance of information items are introduced and an example session demonstrates the prototype. The presence of a spatial model in the user's mind is regarded as an essential contribution towards natural interaction and reduction of cognitive costs during retrieval dialogues.
    Source
    Proceeding CHI '95 Conference Companion on Human Factors in Computing Systems
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
  18. Batorowska, H.; Kaminska-Czubala, B.: Information retrieval support : visualisation of the information space of a document (2014) 0.03
    0.030951615 = product of:
      0.06190323 = sum of:
        0.023615643 = weight(_text_:retrieval in 1444) [ClassicSimilarity], result of:
          0.023615643 = score(doc=1444,freq=4.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.18905719 = fieldWeight in 1444, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.03125 = fieldNorm(doc=1444)
        0.019452432 = weight(_text_:of in 1444) [ClassicSimilarity], result of:
          0.019452432 = score(doc=1444,freq=38.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.30123898 = fieldWeight in 1444, product of:
              6.164414 = tf(freq=38.0), with freq of:
                38.0 = termFreq=38.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=1444)
        0.007645456 = product of:
          0.015290912 = sum of:
            0.015290912 = weight(_text_:on in 1444) [ClassicSimilarity], result of:
              0.015290912 = score(doc=1444,freq=6.0), product of:
                0.090823986 = queryWeight, product of:
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.041294612 = queryNorm
                0.16835764 = fieldWeight in 1444, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1444)
          0.5 = coord(1/2)
        0.0111897 = product of:
          0.0223794 = sum of:
            0.0223794 = weight(_text_:22 in 1444) [ClassicSimilarity], result of:
              0.0223794 = score(doc=1444,freq=2.0), product of:
                0.1446067 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.041294612 = queryNorm
                0.15476047 = fieldWeight in 1444, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1444)
          0.5 = coord(1/2)
      0.5 = coord(4/8)
    
    Abstract
    Acquiring knowledge in any field involves information retrieval, i.e. searching the available documents to identify answers to the queries concerning the selected objects. Knowing the keywords which are names of the objects will enable situating the user's query in the information space organized as a thesaurus or faceted classification. Objectives: Identification the areas in the information space which correspond to gaps in the user's personal knowledge or in the domain knowledge might become useful in theory or practice. The aim of this paper is to present a realistic information-space model of a self-authored full-text document on information culture, indexed by the author of this article. Methodology: Having established the relations between the terms, particular modules (sets of terms connected by relations used in facet classification) are situated on a plain, similarly to a communication map. Conclusions drawn from the "journey" on the map, which is a visualization of the knowledge contained in the analysed document, are the crucial part of this paper. Results: The direct result of the research is the created model of information space visualization of a given document (book, article, website). The proposed procedure can practically be used as a new form of representation in order to map the contents of academic books and articles, beside the traditional index form, especially as an e-book auxiliary tool. In teaching, visualization of the information space of a document can be used to help students understand the issues of: classification, categorization and representation of new knowledge emerging in human mind.
    Source
    Knowledge organization in the 21st century: between historical patterns and future prospects. Proceedings of the Thirteenth International ISKO Conference 19-22 May 2014, Kraków, Poland. Ed.: Wieslaw Babik
  19. Linden, E.J. van der; Vliegen, R.; Wijk, J.J. van: Visual Universal Decimal Classification (2007) 0.03
    0.030484932 = product of:
      0.08129315 = sum of:
        0.020873476 = weight(_text_:retrieval in 548) [ClassicSimilarity], result of:
          0.020873476 = score(doc=548,freq=2.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.16710453 = fieldWeight in 548, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0390625 = fieldNorm(doc=548)
        0.04277933 = weight(_text_:use in 548) [ClassicSimilarity], result of:
          0.04277933 = score(doc=548,freq=8.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.3383162 = fieldWeight in 548, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0390625 = fieldNorm(doc=548)
        0.017640345 = weight(_text_:of in 548) [ClassicSimilarity], result of:
          0.017640345 = score(doc=548,freq=20.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.27317715 = fieldWeight in 548, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=548)
      0.375 = coord(3/8)
    
    Abstract
    UDC aims to be a consistent and complete classification system, that enables practitioners to classify documents swiftly and smoothly. The eventual goal of UDC is to enable the public at large to retrieve documents from large collections of documents that are classified with UDC. The large size of the UDC Master Reference File, MRF with over 66.000 records, makes it difficult to obtain an overview and to understand its structure. Moreover, finding the right classification in MRF turns out to be difficult in practice. Last but not least, retrieval of documents requires insight and understanding of the coding system. Visualization is an effective means to support the development of UDC as well as its use by practitioners. Moreover, visualization offers possibilities to use the classification without use of the coding system as such. MagnaView has developed an application which demonstrates the use of interactive visualization to face these challenges. In our presentation, we discuss these challenges, and we give a demonstration of the way the application helps face these. Examples of visualizations can be found below.
  20. Beagle, D.: Visualizing keyword distribution across multidisciplinary c-space (2003) 0.03
    0.030173076 = product of:
      0.060346153 = sum of:
        0.021692354 = weight(_text_:retrieval in 1202) [ClassicSimilarity], result of:
          0.021692354 = score(doc=1202,freq=6.0), product of:
            0.124912694 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.041294612 = queryNorm
            0.17366013 = fieldWeight in 1202, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1202)
        0.012833798 = weight(_text_:use in 1202) [ClassicSimilarity], result of:
          0.012833798 = score(doc=1202,freq=2.0), product of:
            0.12644777 = queryWeight, product of:
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.041294612 = queryNorm
            0.101494856 = fieldWeight in 1202, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0620887 = idf(docFreq=5623, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1202)
        0.017710768 = weight(_text_:of in 1202) [ClassicSimilarity], result of:
          0.017710768 = score(doc=1202,freq=56.0), product of:
            0.06457475 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.041294612 = queryNorm
            0.2742677 = fieldWeight in 1202, product of:
              7.483315 = tf(freq=56.0), with freq of:
                56.0 = termFreq=56.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1202)
        0.008109231 = product of:
          0.016218461 = sum of:
            0.016218461 = weight(_text_:on in 1202) [ClassicSimilarity], result of:
              0.016218461 = score(doc=1202,freq=12.0), product of:
                0.090823986 = queryWeight, product of:
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.041294612 = queryNorm
                0.17857024 = fieldWeight in 1202, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  2.199415 = idf(docFreq=13325, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1202)
          0.5 = coord(1/2)
      0.5 = coord(4/8)
    
    Abstract
    The concept of c-space is proposed as a visualization schema relating containers of content to cataloging surrogates and classification structures. Possible applications of keyword vector clusters within c-space could include improved retrieval rates through the use of captioning within visual hierarchies, tracings of semantic bleeding among subclasses, and access to buried knowledge within subject-neutral publication containers. The Scholastica Project is described as one example, following a tradition of research dating back to the 1980's. Preliminary focus group assessment indicates that this type of classification rendering may offer digital library searchers enriched entry strategies and an expanded range of re-entry vocabularies. Those of us who work in traditional libraries typically assume that our systems of classification: Library of Congress Classification (LCC) and Dewey Decimal Classification (DDC), are descriptive rather than prescriptive. In other words, LCC classes and subclasses approximate natural groupings of texts that reflect an underlying order of knowledge, rather than arbitrary categories prescribed by librarians to facilitate efficient shelving. Philosophical support for this assumption has traditionally been found in a number of places, from the archetypal tree of knowledge, to Aristotelian categories, to the concept of discursive formations proposed by Michel Foucault. Gary P. Radford has elegantly described an encounter with Foucault's discursive formations in the traditional library setting: "Just by looking at the titles on the spines, you can see how the books cluster together...You can identify those books that seem to form the heart of the discursive formation and those books that reside on the margins. Moving along the shelves, you see those books that tend to bleed over into other classifications and that straddle multiple discursive formations. You can physically and sensually experience...those points that feel like state borders or national boundaries, those points where one subject ends and another begins, or those magical places where one subject has morphed into another..."
    But what happens to this awareness in a digital library? Can discursive formations be represented in cyberspace, perhaps through diagrams in a visualization interface? And would such a schema be helpful to a digital library user? To approach this question, it is worth taking a moment to reconsider what Radford is looking at. First, he looks at titles to see how the books cluster. To illustrate, I scanned one hundred books on the shelves of a college library under subclass HT 101-395, defined by the LCC subclass caption as Urban groups. The City. Urban sociology. Of the first 100 titles in this sequence, fifty included the word "urban" or variants (e.g. "urbanization"). Another thirty-five used the word "city" or variants. These keywords appear to mark their titles as the heart of this discursive formation. The scattering of titles not using "urban" or "city" used related terms such as "town," "community," or in one case "skyscrapers." So we immediately see some empirical correlation between keywords and classification. But we also see a problem with the commonly used search technique of title-keyword. A student interested in urban studies will want to know about this entire subclass, and may wish to browse every title available therein. A title-keyword search on "urban" will retrieve only half of the titles, while a search on "city" will retrieve just over a third. There will be no overlap, since no titles in this sample contain both words. The only place where both words appear in a common string is in the LCC subclass caption, but captions are not typically indexed in library Online Public Access Catalogs (OPACs). In a traditional library, this problem is mitigated when the student goes to the shelf looking for any one of the books and suddenly discovers a much wider selection than the keyword search had led him to expect. But in a digital library, the issue of non-retrieval can be more problematic, as studies have indicated. Micco and Popp reported that, in a study funded partly by the U.S. Department of Education, 65 of 73 unskilled users searching for material on U.S./Soviet foreign relations found some material but never realized they had missed a large percentage of what was in the database.
    Theme
    Klassifikationssysteme im Online-Retrieval

Years

Languages

  • e 143
  • d 19
  • a 1
  • More… Less…

Types

  • a 133
  • el 28
  • m 15
  • x 9
  • s 3
  • r 2
  • b 1
  • p 1
  • More… Less…

Subjects

Classifications