Search (30 results, page 1 of 2)

  • × classification_ss:"06.74 / Informationssysteme"
  1. Survey of text mining : clustering, classification, and retrieval (2004) 0.05
    0.049204174 = product of:
      0.09840835 = sum of:
        0.06843241 = weight(_text_:data in 804) [ClassicSimilarity], result of:
          0.06843241 = score(doc=804,freq=14.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.46216056 = fieldWeight in 804, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0390625 = fieldNorm(doc=804)
        0.029975938 = product of:
          0.059951875 = sum of:
            0.059951875 = weight(_text_:processing in 804) [ClassicSimilarity], result of:
              0.059951875 = score(doc=804,freq=4.0), product of:
                0.18956426 = queryWeight, product of:
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.046827413 = queryNorm
                0.3162615 = fieldWeight in 804, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=804)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Extracting content from text continues to be an important research problem for information processing and management. Approaches to capture the semantics of text-based document collections may be based on Bayesian models, probability theory, vector space models, statistical models, or even graph theory. As the volume of digitized textual media continues to grow, so does the need for designing robust, scalable indexing and search strategies (software) to meet a variety of user needs. Knowledge extraction or creation from text requires systematic yet reliable processing that can be codified and adapted for changing needs and environments. This book will draw upon experts in both academia and industry to recommend practical approaches to the purification, indexing, and mining of textual information. It will address document identification, clustering and categorizing documents, cleaning text, and visualizing semantic models of text.
    Classification
    ST 270 Informatik / Monographien / Software und -entwicklung / Datenbanken, Datenbanksysteme, Data base management, Informationssysteme
    LCSH
    Data mining ; Information retrieval
    Data mining / Congresses (GBV)
    RVK
    ST 270 Informatik / Monographien / Software und -entwicklung / Datenbanken, Datenbanksysteme, Data base management, Informationssysteme
    Subject
    Data mining ; Information retrieval
    Data mining / Congresses (GBV)
    Theme
    Data Mining
  2. Information visualization in data mining and knowledge discovery (2002) 0.04
    0.039750893 = product of:
      0.079501785 = sum of:
        0.07315732 = weight(_text_:data in 1789) [ClassicSimilarity], result of:
          0.07315732 = score(doc=1789,freq=100.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.49407038 = fieldWeight in 1789, product of:
              10.0 = tf(freq=100.0), with freq of:
                100.0 = termFreq=100.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.015625 = fieldNorm(doc=1789)
        0.006344468 = product of:
          0.012688936 = sum of:
            0.012688936 = weight(_text_:22 in 1789) [ClassicSimilarity], result of:
              0.012688936 = score(doc=1789,freq=2.0), product of:
                0.16398162 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046827413 = queryNorm
                0.07738023 = fieldWeight in 1789, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.015625 = fieldNorm(doc=1789)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Date
    23. 3.2008 19:10:22
    Footnote
    Rez. in: JASIST 54(2003) no.9, S.905-906 (C.A. Badurek): "Visual approaches for knowledge discovery in very large databases are a prime research need for information scientists focused an extracting meaningful information from the ever growing stores of data from a variety of domains, including business, the geosciences, and satellite and medical imagery. This work presents a summary of research efforts in the fields of data mining, knowledge discovery, and data visualization with the goal of aiding the integration of research approaches and techniques from these major fields. The editors, leading computer scientists from academia and industry, present a collection of 32 papers from contributors who are incorporating visualization and data mining techniques through academic research as well application development in industry and government agencies. Information Visualization focuses upon techniques to enhance the natural abilities of humans to visually understand data, in particular, large-scale data sets. It is primarily concerned with developing interactive graphical representations to enable users to more intuitively make sense of multidimensional data as part of the data exploration process. It includes research from computer science, psychology, human-computer interaction, statistics, and information science. Knowledge Discovery in Databases (KDD) most often refers to the process of mining databases for previously unknown patterns and trends in data. Data mining refers to the particular computational methods or algorithms used in this process. The data mining research field is most related to computational advances in database theory, artificial intelligence and machine learning. This work compiles research summaries from these main research areas in order to provide "a reference work containing the collection of thoughts and ideas of noted researchers from the fields of data mining and data visualization" (p. 8). It addresses these areas in three main sections: the first an data visualization, the second an KDD and model visualization, and the last an using visualization in the knowledge discovery process. The seven chapters of Part One focus upon methodologies and successful techniques from the field of Data Visualization. Hoffman and Grinstein (Chapter 2) give a particularly good overview of the field of data visualization and its potential application to data mining. An introduction to the terminology of data visualization, relation to perceptual and cognitive science, and discussion of the major visualization display techniques are presented. Discussion and illustration explain the usefulness and proper context of such data visualization techniques as scatter plots, 2D and 3D isosurfaces, glyphs, parallel coordinates, and radial coordinate visualizations. Remaining chapters present the need for standardization of visualization methods, discussion of user requirements in the development of tools, and examples of using information visualization in addressing research problems.
    In 13 chapters, Part Two provides an introduction to KDD, an overview of data mining techniques, and examples of the usefulness of data model visualizations. The importance of visualization throughout the KDD process is stressed in many of the chapters. In particular, the need for measures of visualization effectiveness, benchmarking for identifying best practices, and the use of standardized sample data sets is convincingly presented. Many of the important data mining approaches are discussed in this complementary context. Cluster and outlier detection, classification techniques, and rule discovery algorithms are presented as the basic techniques common to the KDD process. The potential effectiveness of using visualization in the data modeling process are illustrated in chapters focused an using visualization for helping users understand the KDD process, ask questions and form hypotheses about their data, and evaluate the accuracy and veracity of their results. The 11 chapters of Part Three provide an overview of the KDD process and successful approaches to integrating KDD, data mining, and visualization in complementary domains. Rhodes (Chapter 21) begins this section with an excellent overview of the relation between the KDD process and data mining techniques. He states that the "primary goals of data mining are to describe the existing data and to predict the behavior or characteristics of future data of the same type" (p. 281). These goals are met by data mining tasks such as classification, regression, clustering, summarization, dependency modeling, and change or deviation detection. Subsequent chapters demonstrate how visualization can aid users in the interactive process of knowledge discovery by graphically representing the results from these iterative tasks. Finally, examples of the usefulness of integrating visualization and data mining tools in the domain of business, imagery and text mining, and massive data sets are provided. This text concludes with a thorough and useful 17-page index and lengthy yet integrating 17-page summary of the academic and industrial backgrounds of the contributing authors. A 16-page set of color inserts provide a better representation of the visualizations discussed, and a URL provided suggests that readers may view all the book's figures in color on-line, although as of this submission date it only provides access to a summary of the book and its contents. The overall contribution of this work is its focus an bridging two distinct areas of research, making it a valuable addition to the Morgan Kaufmann Series in Database Management Systems. The editors of this text have met their main goal of providing the first textbook integrating knowledge discovery, data mining, and visualization. Although it contributes greatly to our under- standing of the development and current state of the field, a major weakness of this text is that there is no concluding chapter to discuss the contributions of the sum of these contributed papers or give direction to possible future areas of research. "Integration of expertise between two different disciplines is a difficult process of communication and reeducation. Integrating data mining and visualization is particularly complex because each of these fields in itself must draw an a wide range of research experience" (p. 300). Although this work contributes to the crossdisciplinary communication needed to advance visualization in KDD, a more formal call for an interdisciplinary research agenda in a concluding chapter would have provided a more satisfying conclusion to a very good introductory text.
    With contributors almost exclusively from the computer science field, the intended audience of this work is heavily slanted towards a computer science perspective. However, it is highly readable and provides introductory material that would be useful to information scientists from a variety of domains. Yet, much interesting work in information visualization from other fields could have been included giving the work more of an interdisciplinary perspective to complement their goals of integrating work in this area. Unfortunately, many of the application chapters are these, shallow, and lack complementary illustrations of visualization techniques or user interfaces used. However, they do provide insight into the many applications being developed in this rapidly expanding field. The authors have successfully put together a highly useful reference text for the data mining and information visualization communities. Those interested in a good introduction and overview of complementary research areas in these fields will be satisfied with this collection of papers. The focus upon integrating data visualization with data mining complements texts in each of these fields, such as Advances in Knowledge Discovery and Data Mining (Fayyad et al., MIT Press) and Readings in Information Visualization: Using Vision to Think (Card et. al., Morgan Kauffman). This unique work is a good starting point for future interaction between researchers in the fields of data visualization and data mining and makes a good accompaniment for a course focused an integrating these areas or to the main reference texts in these fields."
    LCSH
    Data mining
    RSWK
    Visualisierung / Computergraphik / Data Mining
    Data Mining / Visualisierung / Aufsatzsammlung (BVB)
    Series
    Morgan Kaufmann series in data management systems
    Subject
    Visualisierung / Computergraphik / Data Mining
    Data Mining / Visualisierung / Aufsatzsammlung (BVB)
    Data mining
    Theme
    Data Mining
  3. Berry, M.W.; Browne, M.: Understanding search engines : mathematical modeling and text retrieval (1999) 0.03
    0.033504575 = product of:
      0.06700915 = sum of:
        0.031038022 = weight(_text_:data in 5777) [ClassicSimilarity], result of:
          0.031038022 = score(doc=5777,freq=2.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.2096163 = fieldWeight in 5777, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046875 = fieldNorm(doc=5777)
        0.035971127 = product of:
          0.071942255 = sum of:
            0.071942255 = weight(_text_:processing in 5777) [ClassicSimilarity], result of:
              0.071942255 = score(doc=5777,freq=4.0), product of:
                0.18956426 = queryWeight, product of:
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.046827413 = queryNorm
                0.3795138 = fieldWeight in 5777, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5777)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    This book discusses many of the key design issues for building search engines and emphazises the important role that applied mathematics can play in improving information retrieval. The authors discuss not only important data structures, algorithms, and software but also user-centered issues such as interfaces, manual indexing, and document preparation. They also present some of the current problems in information retrieval that many not be familiar to applied mathematicians and computer scientists and some of the driving computational methods (SVD, SDD) for automated conceptual indexing
    LCSH
    Text processing (Computer science)
    Subject
    Text processing (Computer science)
  4. Grossman, D.A.; Frieder, O.: Information retrieval : algorithms and heuristics (2004) 0.03
    0.029316615 = product of:
      0.05863323 = sum of:
        0.029262928 = weight(_text_:data in 1486) [ClassicSimilarity], result of:
          0.029262928 = score(doc=1486,freq=4.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.19762816 = fieldWeight in 1486, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.03125 = fieldNorm(doc=1486)
        0.029370302 = product of:
          0.058740605 = sum of:
            0.058740605 = weight(_text_:processing in 1486) [ClassicSimilarity], result of:
              0.058740605 = score(doc=1486,freq=6.0), product of:
                0.18956426 = queryWeight, product of:
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.046827413 = queryNorm
                0.30987173 = fieldWeight in 1486, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1486)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Interested in how an efficient search engine works? Want to know what algorithms are used to rank resulting documents in response to user requests? The authors answer these and other key information on retrieval design and implementation questions is provided. This book is not yet another high level text. Instead, algorithms are thoroughly described, making this book ideally suited for both computer science students and practitioners who work on search-related applications. As stated in the foreword, this book provides a current, broad, and detailed overview of the field and is the only one that does so. Examples are used throughout to illustrate the algorithms. The authors explain how a query is ranked against a document collection using either a single or a combination of retrieval strategies, and how an assortment of utilities are integrated into the query processing scheme to improve these rankings. Methods for building and compressing text indexes, querying and retrieving documents in multiple languages, and using parallel or distributed processing to expedite the search are likewise described. This edition is a major expansion of the one published in 1998. Neuaufl. 2005: Besides updating the entire book with current techniques, it includes new sections on language models, cross-language information retrieval, peer-to-peer processing, XML search, mediators, and duplicate document detection.
    Classification
    ST 270 Informatik / Monographien / Software und -entwicklung / Datenbanken, Datenbanksysteme, Data base management, Informationssysteme
    RVK
    ST 270 Informatik / Monographien / Software und -entwicklung / Datenbanken, Datenbanksysteme, Data base management, Informationssysteme
  5. Berry, M.W.; Browne, M.: Understanding search engines : mathematical modeling and text retrieval (2005) 0.02
    0.022336382 = product of:
      0.044672765 = sum of:
        0.020692015 = weight(_text_:data in 7) [ClassicSimilarity], result of:
          0.020692015 = score(doc=7,freq=2.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.1397442 = fieldWeight in 7, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.03125 = fieldNorm(doc=7)
        0.02398075 = product of:
          0.0479615 = sum of:
            0.0479615 = weight(_text_:processing in 7) [ClassicSimilarity], result of:
              0.0479615 = score(doc=7,freq=4.0), product of:
                0.18956426 = queryWeight, product of:
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.046827413 = queryNorm
                0.2530092 = fieldWeight in 7, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.03125 = fieldNorm(doc=7)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    The second edition of Understanding Search Engines: Mathematical Modeling and Text Retrieval follows the basic premise of the first edition by discussing many of the key design issues for building search engines and emphasizing the important role that applied mathematics can play in improving information retrieval. The authors discuss important data structures, algorithms, and software as well as user-centered issues such as interfaces, manual indexing, and document preparation. Significant changes bring the text up to date on current information retrieval methods: for example the addition of a new chapter on link-structure algorithms used in search engines such as Google. The chapter on user interface has been rewritten to specifically focus on search engine usability. In addition the authors have added new recommendations for further reading and expanded the bibliography, and have updated and streamlined the index to make it more reader friendly.
    LCSH
    Text processing (Computer science)
    Subject
    Text processing (Computer science)
  6. Langville, A.N.; Meyer, C.D.: Google's PageRank and beyond : the science of search engine rankings (2006) 0.02
    0.01996638 = product of:
      0.03993276 = sum of:
        0.021947198 = weight(_text_:data in 6) [ClassicSimilarity], result of:
          0.021947198 = score(doc=6,freq=4.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.14822112 = fieldWeight in 6, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0234375 = fieldNorm(doc=6)
        0.017985564 = product of:
          0.035971127 = sum of:
            0.035971127 = weight(_text_:processing in 6) [ClassicSimilarity], result of:
              0.035971127 = score(doc=6,freq=4.0), product of:
                0.18956426 = queryWeight, product of:
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.046827413 = queryNorm
                0.1897569 = fieldWeight in 6, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=6)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Why doesn't your home page appear on the first page of search results, even when you query your own name? How do other Web pages always appear at the top? What creates these powerful rankings? And how? The first book ever about the science of Web page rankings, "Google's PageRank and Beyond" supplies the answers to these and other questions and more. The book serves two very different audiences: the curious science reader and the technical computational reader. The chapters build in mathematical sophistication, so that the first five are accessible to the general academic reader. While other chapters are much more mathematical in nature, each one contains something for both audiences. For example, the authors include entertaining asides such as how search engines make money and how the Great Firewall of China influences research. The book includes an extensive background chapter designed to help readers learn more about the mathematics of search engines, and it contains several MATLAB codes and links to sample Web data sets. The philosophy throughout is to encourage readers to experiment with the ideas and algorithms in the text. Any business seriously interested in improving its rankings in the major search engines can benefit from the clear examples, sample code, and list of resources provided. It includes: many illustrative examples and entertaining asides; MATLAB code; accessible and informal style; and complete and self-contained section for mathematics review.
    Content
    Inhalt: Chapter 1. Introduction to Web Search Engines: 1.1 A Short History of Information Retrieval - 1.2 An Overview of Traditional Information Retrieval - 1.3 Web Information Retrieval Chapter 2. Crawling, Indexing, and Query Processing: 2.1 Crawling - 2.2 The Content Index - 2.3 Query Processing Chapter 3. Ranking Webpages by Popularity: 3.1 The Scene in 1998 - 3.2 Two Theses - 3.3 Query-Independence Chapter 4. The Mathematics of Google's PageRank: 4.1 The Original Summation Formula for PageRank - 4.2 Matrix Representation of the Summation Equations - 4.3 Problems with the Iterative Process - 4.4 A Little Markov Chain Theory - 4.5 Early Adjustments to the Basic Model - 4.6 Computation of the PageRank Vector - 4.7 Theorem and Proof for Spectrum of the Google Matrix Chapter 5. Parameters in the PageRank Model: 5.1 The a Factor - 5.2 The Hyperlink Matrix H - 5.3 The Teleportation Matrix E Chapter 6. The Sensitivity of PageRank; 6.1 Sensitivity with respect to alpha - 6.2 Sensitivity with respect to H - 6.3 Sensitivity with respect to vT - 6.4 Other Analyses of Sensitivity - 6.5 Sensitivity Theorems and Proofs Chapter 7. The PageRank Problem as a Linear System: 7.1 Properties of (I - alphaS) - 7.2 Properties of (I - alphaH) - 7.3 Proof of the PageRank Sparse Linear System Chapter 8. Issues in Large-Scale Implementation of PageRank: 8.1 Storage Issues - 8.2 Convergence Criterion - 8.3 Accuracy - 8.4 Dangling Nodes - 8.5 Back Button Modeling
    Chapter 9. Accelerating the Computation of PageRank: 9.1 An Adaptive Power Method - 9.2 Extrapolation - 9.3 Aggregation - 9.4 Other Numerical Methods Chapter 10. Updating the PageRank Vector: 10.1 The Two Updating Problems and their History - 10.2 Restarting the Power Method - 10.3 Approximate Updating Using Approximate Aggregation - 10.4 Exact Aggregation - 10.5 Exact vs. Approximate Aggregation - 10.6 Updating with Iterative Aggregation - 10.7 Determining the Partition - 10.8 Conclusions Chapter 11. The HITS Method for Ranking Webpages: 11.1 The HITS Algorithm - 11.2 HITS Implementation - 11.3 HITS Convergence - 11.4 HITS Example - 11.5 Strengths and Weaknesses of HITS - 11.6 HITS's Relationship to Bibliometrics - 11.7 Query-Independent HITS - 11.8 Accelerating HITS - 11.9 HITS Sensitivity Chapter 12. Other Link Methods for Ranking Webpages: 12.1 SALSA - 12.2 Hybrid Ranking Methods - 12.3 Rankings based on Traffic Flow Chapter 13. The Future of Web Information Retrieval: 13.1 Spam - 13.2 Personalization - 13.3 Clustering - 13.4 Intelligent Agents - 13.5 Trends and Time-Sensitive Search - 13.6 Privacy and Censorship - 13.7 Library Classification Schemes - 13.8 Data Fusion Chapter 14. Resources for Web Information Retrieval: 14.1 Resources for Getting Started - 14.2 Resources for Serious Study Chapter 15. The Mathematics Guide: 15.1 Linear Algebra - 15.2 Perron-Frobenius Theory - 15.3 Markov Chains - 15.4 Perron Complementation - 15.5 Stochastic Complementation - 15.6 Censoring - 15.7 Aggregation - 15.8 Disaggregation
  7. Spinning the Semantic Web : bringing the World Wide Web to its full potential (2003) 0.02
    0.016471423 = product of:
      0.032942846 = sum of:
        0.018105512 = weight(_text_:data in 1981) [ClassicSimilarity], result of:
          0.018105512 = score(doc=1981,freq=2.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.12227618 = fieldWeight in 1981, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1981)
        0.014837332 = product of:
          0.029674664 = sum of:
            0.029674664 = weight(_text_:processing in 1981) [ClassicSimilarity], result of:
              0.029674664 = score(doc=1981,freq=2.0), product of:
                0.18956426 = queryWeight, product of:
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.046827413 = queryNorm
                0.15654145 = fieldWeight in 1981, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1981)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    As the World Wide Web continues to expand, it becomes increasingly difficult for users to obtain information efficiently. Because most search engines read format languages such as HTML or SGML, search results reflect formatting tags more than actual page content, which is expressed in natural language. Spinning the Semantic Web describes an exciting new type of hierarchy and standardization that will replace the current "Web of links" with a "Web of meaning." Using a flexible set of languages and tools, the Semantic Web will make all available information - display elements, metadata, services, images, and especially content - accessible. The result will be an immense repository of information accessible for a wide range of new applications. This first handbook for the Semantic Web covers, among other topics, software agents that can negotiate and collect information, markup languages that can tag many more types of information in a document, and knowledge systems that enable machines to read Web pages and determine their reliability. The truly interdisciplinary Semantic Web combines aspects of artificial intelligence, markup languages, natural language processing, information retrieval, knowledge representation, intelligent agents, and databases.
    Content
    Inhalt: Tim Bemers-Lee: The Original Dream - Re-enter Machines - Where Are We Now? - The World Wide Web Consortium - Where Is the Web Going Next? / Dieter Fensel, James Hendler, Henry Lieberman, and Wolfgang Wahlster: Why Is There a Need for the Semantic Web and What Will It Provide? - How the Semantic Web Will Be Possible / Jeff Heflin, James Hendler, and Sean Luke: SHOE: A Blueprint for the Semantic Web / Deborah L. McGuinness, Richard Fikes, Lynn Andrea Stein, and James Hendler: DAML-ONT: An Ontology Language for the Semantic Web / Michel Klein, Jeen Broekstra, Dieter Fensel, Frank van Harmelen, and Ian Horrocks: Ontologies and Schema Languages on the Web / Borys Omelayenko, Monica Crubezy, Dieter Fensel, Richard Benjamins, Bob Wielinga, Enrico Motta, Mark Musen, and Ying Ding: UPML: The Language and Tool Support for Making the Semantic Web Alive / Deborah L. McGuinness: Ontologies Come of Age / Jeen Broekstra, Arjohn Kampman, and Frank van Harmelen: Sesame: An Architecture for Storing and Querying RDF Data and Schema Information / Rob Jasper and Mike Uschold: Enabling Task-Centered Knowledge Support through Semantic Markup / Yolanda Gil: Knowledge Mobility: Semantics for the Web as a White Knight for Knowledge-Based Systems / Sanjeev Thacker, Amit Sheth, and Shuchi Patel: Complex Relationships for the Semantic Web / Alexander Maedche, Steffen Staab, Nenad Stojanovic, Rudi Studer, and York Sure: SEmantic portAL: The SEAL Approach / Ora Lassila and Mark Adler: Semantic Gadgets: Ubiquitous Computing Meets the Semantic Web / Christopher Frye, Mike Plusch, and Henry Lieberman: Static and Dynamic Semantics of the Web / Masahiro Hori: Semantic Annotation for Web Content Adaptation / Austin Tate, Jeff Dalton, John Levine, and Alex Nixon: Task-Achieving Agents on the World Wide Web
  8. Multimedia content and the Semantic Web : methods, standards, and tools (2005) 0.01
    0.012167135 = product of:
      0.04866854 = sum of:
        0.04866854 = sum of:
          0.021196188 = weight(_text_:processing in 150) [ClassicSimilarity], result of:
            0.021196188 = score(doc=150,freq=2.0), product of:
              0.18956426 = queryWeight, product of:
                4.048147 = idf(docFreq=2097, maxDocs=44218)
                0.046827413 = queryNorm
              0.111815326 = fieldWeight in 150, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.048147 = idf(docFreq=2097, maxDocs=44218)
                0.01953125 = fieldNorm(doc=150)
          0.027472353 = weight(_text_:22 in 150) [ClassicSimilarity], result of:
            0.027472353 = score(doc=150,freq=6.0), product of:
              0.16398162 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046827413 = queryNorm
              0.16753313 = fieldWeight in 150, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.01953125 = fieldNorm(doc=150)
      0.25 = coord(1/4)
    
    Classification
    006.7 22
    Date
    7. 3.2007 19:30:22
    DDC
    006.7 22
    Footnote
    The final part of the book discusses research in multimedia content management systems and the semantic web, and presents examples and applications for semantic multimedia analysis in search and retrieval systems. These chapters describe example systems in which current projects have been implemented, and include extensive results and real demonstrations. For example, real case scenarios such as ECommerce medical applications and Web services have been introduced. Topics in natural language, speech and image processing techniques and their application for multimedia indexing, and content-based retrieval have been elaborated upon with extensive examples and deployment methods. The editors of the book themselves provide the readers with a chapter about their latest research results on knowledge-based multimedia content indexing and retrieval. Some interesting applications for multimedia content and the semantic web are introduced. Applications that have taken advantage of the metadata provided by MPEG7 in order to realize advance-access services for multimedia content have been provided. The applications discussed in the third part of the book provide useful guidance to researchers and practitioners properly planning to implement semantic multimedia analysis techniques in new research and development projects in both academia and industry. A fourth part should be added to this book: performance measurements for integrated approaches of multimedia analysis and the semantic web. Performance of the semantic approach is a very sophisticated issue and requires extensive elaboration and effort. Measuring the semantic search is an ongoing research area; several chapters concerning performance measurement and analysis would be required to adequately cover this area and introduce it to readers."
  9. Gebhardt, F.: Dokumentationssysteme (1981) 0.01
    0.010973599 = product of:
      0.043894395 = sum of:
        0.043894395 = weight(_text_:data in 1560) [ClassicSimilarity], result of:
          0.043894395 = score(doc=1560,freq=4.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.29644224 = fieldWeight in 1560, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046875 = fieldNorm(doc=1560)
      0.25 = coord(1/4)
    
    Classification
    ST 271 Informatik / Monographien / Software und -entwicklung / Datenbanken, Datenbanksysteme, Data base management, Informationssysteme / Einzelne Datenbanksprachen und Datenbanksysteme
    RVK
    ST 271 Informatik / Monographien / Software und -entwicklung / Datenbanken, Datenbanksysteme, Data base management, Informationssysteme / Einzelne Datenbanksprachen und Datenbanksysteme
  10. Henzler, R.G.: Information und Dokumentation : Sammeln, Speichern und Wiedergewinnen von Fachinformation in Datenbanken (1992) 0.01
    0.0103460075 = product of:
      0.04138403 = sum of:
        0.04138403 = weight(_text_:data in 4839) [ClassicSimilarity], result of:
          0.04138403 = score(doc=4839,freq=8.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.2794884 = fieldWeight in 4839, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.03125 = fieldNorm(doc=4839)
      0.25 = coord(1/4)
    
    Classification
    ST 271 Informatik / Monographien / Software und -entwicklung / Datenbanken, Datenbanksysteme, Data base management, Informationssysteme / Einzelne Datenbanksprachen und Datenbanksysteme
    ST 270 Informatik / Monographien / Software und -entwicklung / Datenbanken, Datenbanksysteme, Data base management, Informationssysteme
    RVK
    ST 271 Informatik / Monographien / Software und -entwicklung / Datenbanken, Datenbanksysteme, Data base management, Informationssysteme / Einzelne Datenbanksprachen und Datenbanksysteme
    ST 270 Informatik / Monographien / Software und -entwicklung / Datenbanken, Datenbanksysteme, Data base management, Informationssysteme
  11. Oßwald, A.: Dokumentlieferung im Zeitalter elektronischen Publizierens (1992) 0.01
    0.009144665 = product of:
      0.03657866 = sum of:
        0.03657866 = weight(_text_:data in 1603) [ClassicSimilarity], result of:
          0.03657866 = score(doc=1603,freq=4.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.24703519 = fieldWeight in 1603, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1603)
      0.25 = coord(1/4)
    
    Classification
    ST 271 Informatik / Monographien / Software und -entwicklung / Datenbanken, Datenbanksysteme, Data base management, Informationssysteme / Einzelne Datenbanksprachen und Datenbanksysteme
    RVK
    ST 271 Informatik / Monographien / Software und -entwicklung / Datenbanken, Datenbanksysteme, Data base management, Informationssysteme / Einzelne Datenbanksprachen und Datenbanksysteme
  12. Hare, C.E.; McLeod, J.: How to manage records in the e-environment : 2nd ed. (2006) 0.01
    0.009052756 = product of:
      0.036211025 = sum of:
        0.036211025 = weight(_text_:data in 1749) [ClassicSimilarity], result of:
          0.036211025 = score(doc=1749,freq=2.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.24455236 = fieldWeight in 1749, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1749)
      0.25 = coord(1/4)
    
    Abstract
    A practical approach to developing and operating an effective programme to manage hybrid records within an organization. This title positions records management as an integral business function linked to the organisation's business aims and objectives. The authors also address the records requirements of new and significant pieces of legislation, such as data protection and freedom of information, as well as exploring strategies for managing electronic records. Bullet points, checklists and examples assist the reader throughout, making this a one-stop resource for information in this area.
  13. Glöckner-Rist, A.: Suchfragen im Information Retrieval : eine empirische Untersuchung zum Rechercheverhalten von Informationsvermittlern und Endbenutzern (1992) 0.01
    0.007315732 = product of:
      0.029262928 = sum of:
        0.029262928 = weight(_text_:data in 1602) [ClassicSimilarity], result of:
          0.029262928 = score(doc=1602,freq=4.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.19762816 = fieldWeight in 1602, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.03125 = fieldNorm(doc=1602)
      0.25 = coord(1/4)
    
    Classification
    ST 270 Informatik / Monographien / Software und -entwicklung / Datenbanken, Datenbanksysteme, Data base management, Informationssysteme
    RVK
    ST 270 Informatik / Monographien / Software und -entwicklung / Datenbanken, Datenbanksysteme, Data base management, Informationssysteme
  14. Lavrenko, V.: ¬A generative theory of relevance (2009) 0.01
    0.006466255 = product of:
      0.02586502 = sum of:
        0.02586502 = weight(_text_:data in 3306) [ClassicSimilarity], result of:
          0.02586502 = score(doc=3306,freq=2.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.17468026 = fieldWeight in 3306, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3306)
      0.25 = coord(1/4)
    
    Abstract
    A modern information retrieval system must have the capability to find, organize and present very different manifestations of information - such as text, pictures, videos or database records - any of which may be of relevance to the user. However, the concept of relevance, while seemingly intuitive, is actually hard to define, and it's even harder to model in a formal way. Lavrenko does not attempt to bring forth a new definition of relevance, nor provide arguments as to why any particular definition might be theoretically superior or more complete. Instead, he takes a widely accepted, albeit somewhat conservative definition, makes several assumptions, and from them develops a new probabilistic model that explicitly captures that notion of relevance. With this book, he makes two major contributions to the field of information retrieval: first, a new way to look at topical relevance, complementing the two dominant models, i.e., the classical probabilistic model and the language modeling approach, and which explicitly combines documents, queries, and relevance in a single formalism; second, a new method for modeling exchangeable sequences of discrete random variables which does not make any structural assumptions about the data and which can also handle rare events. Thus his book is of major interest to researchers and graduate students in information retrieval who specialize in relevance modeling, ranking algorithms, and language modeling.
  15. Levy, S.: In the plex : how Google thinks, works, and shapes our lives (2011) 0.01
    0.006401266 = product of:
      0.025605064 = sum of:
        0.025605064 = weight(_text_:data in 9) [ClassicSimilarity], result of:
          0.025605064 = score(doc=9,freq=4.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.17292464 = fieldWeight in 9, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.02734375 = fieldNorm(doc=9)
      0.25 = coord(1/4)
    
    Abstract
    Few companies in history have ever been as successful and as admired as Google, the company that has transformed the Internet and become an indispensable part of our lives. How has Google done it? Veteran technology reporter Steven Levy was granted unprecedented access to the company, and in this revelatory book he takes readers inside Google headquarters-the Googleplex-to show how Google works. While they were still students at Stanford, Google cofounders Larry Page and Sergey Brin revolutionized Internet search. They followed this brilliant innovation with another, as two of Google's earliest employees found a way to do what no one else had: make billions of dollars from Internet advertising. With this cash cow (until Google's IPO nobody other than Google management had any idea how lucrative the company's ad business was), Google was able to expand dramatically and take on other transformative projects: more efficient data centers, open-source cell phones, free Internet video (YouTube), cloud computing, digitizing books, and much more. The key to Google's success in all these businesses, Levy reveals, is its engineering mind-set and adoption of such Internet values as speed, openness, experimentation, and risk taking. After its unapologetically elitist approach to hiring, Google pampers its engineers-free food and dry cleaning, on-site doctors and masseuses-and gives them all the resources they need to succeed. Even today, with a workforce of more than 23,000, Larry Page signs off on every hire. But has Google lost its innovative edge? It stumbled badly in China-Levy discloses what went wrong and how Brin disagreed with his peers on the China strategy-and now with its newest initiative, social networking, Google is chasing a successful competitor for the first time. Some employees are leaving the company for smaller, nimbler start-ups. Can the company that famously decided not to be evil still compete? No other book has ever turned Google inside out as Levy does with In the Plex.
    Content
    The world according to Google: biography of a search engine -- Googlenomics: cracking the code on internet profits -- Don't be evil: how Google built its culture -- Google's cloud: how Google built data centers and killed the hard drive -- Outside the box: the Google phone company. and the Google t.v. company -- Guge: Google moral dilemma in China -- Google.gov: is what's good for Google, good for government or the public? -- Epilogue: chasing tail lights: trying to crack the social code.
  16. Proceedings of the Second ACM/IEEE-CS Joint Conference on Digital Libraries : July 14 - 18, 2002, Portland, Oregon, USA. (2002) 0.01
    0.0063356096 = product of:
      0.025342438 = sum of:
        0.025342438 = weight(_text_:data in 172) [ClassicSimilarity], result of:
          0.025342438 = score(doc=172,freq=12.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.171151 = fieldWeight in 172, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.015625 = fieldNorm(doc=172)
      0.25 = coord(1/4)
    
    Content
    Inhalt: SESSION: Building and using cultural digital libraries Primarily history: historians and the search for primary source materials (Helen R. Tibbo) - Using the Gamera framework for the recognition of cultural heritage materials (Michael Droettboom, Ichiro Fujinaga, Karl MacMillan, G. Sayeed Chouhury, Tim DiLauro, Mark Patton, Teal Anderson) - Supporting access to large digital oral history archives (Samuel Gustman, Dagobert Soergel, Douglas Oard, William Byrne, Michael Picheny, Bhuvana Ramabhadran, Douglas Greenberg) SESSION: Summarization and question answering Using sentence-selection heuristics to rank text segments in TXTRACTOR (Daniel McDonald, Hsinchun Chen) - Using librarian techniques in automatic text summarization for information retrieval (Min-Yen Kan, Judith L. Klavans) - QuASM: a system for question answering using semi-structured data (David Pinto, Michael Branstein, Ryan Coleman, W. Bruce Croft, Matthew King, Wei Li, Xing Wei) SESSION: Studying users Reading-in-the-small: a study of reading on small form factor devices (Catherine C. Marshall, Christine Ruotolo) - A graph-based recommender system for digital library (Zan Huang, Wingyan Chung, Thian-Huat Ong, Hsinchun Chen) - The effects of topic familiarity on information search behavior (Diane Kelly, Colleen Cool) SESSION: Classification and browsing A language modelling approach to relevance profiling for document browsing (David J. Harper, Sara Coulthard, Sun Yixing) - Compound descriptors in context: a matching function for classifications and thesauri (Douglas Tudhope, Ceri Binding, Dorothee Blocks, Daniel Cunliffe) - Structuring keyword-based queries for web databases (Rodrigo C. Vieira, Pavel Calado, Altigran S. da Silva, Alberto H. F. Laender, Berthier A. Ribeiro-Neto) - An approach to automatic classification of text for information retrieval (Hong Cui, P. Bryan Heidorn, Hong Zhang)
    SESSION: A digital libraries for education Middle school children's use of the ARTEMIS digital library (June Abbas, Cathleen Norris, Elliott Soloway) - Partnership reviewing: a cooperative approach for peer review of complex educational resources (John Weatherley, Tamara Sumner, Michael Khoo, Michael Wright, Marcel Hoffmann) - A digital library for geography examination resources (Lian-Heong Chua, Dion Hoe-Lian Goh, Ee-Peng Lim, Zehua Liu, Rebecca Pei-Hui Ang) - Digital library services for authors of learning materials (Flora McMartin, Youki Terada) SESSION: Novel search environments Integration of simultaneous searching and reference linking across bibliographic resources on the web (William H. Mischo, Thomas G. Habing, Timothy W. Cole) - Exploring discussion lists: steps and directions (Paula S. Newman) - Comparison of two approaches to building a vertical search tool: a case study in the nanotechnology domain (Michael Chau, Hsinchun Chen, Jialun Qin, Yilu Zhou, Yi Qin, Wai-Ki Sung, Daniel McDonald) SESSION: Video and multimedia digital libraries A multilingual, multimodal digital video library system (Michael R. Lyu, Edward Yau, Sam Sze) - A digital library data model for music (Natalia Minibayeva, Jon W. Dunn) - Video-cuebik: adapting image search to video shots (Alexander G. Hauptmann, Norman D. Papernick) - Virtual multimedia libraries built from the web (Neil C. Rowe) - Multi-modal information retrieval from broadcast video using OCR and speech recognition (Alexander G. Hauptmann, Rong Jin, Tobun Dorbin Ng) SESSION: OAI application Extending SDARTS: extracting metadata from web databases and interfacing with the open archives initiative (Panagiotis G. Ipeirotis, Tom Barry, Luis Gravano) - Using the open archives initiative protocols with EAD (Christopher J. Prom, Thomas G. Habing) - Preservation and transition of NCSTRL using an OAI-based architecture (H. Anan, X. Liu, K. Maly, M. Nelson, M. Zubair, J. C. French, E. Fox, P. Shivakumar) - Integrating harvesting into digital library content (David A. Smith, Anne Mahoney, Gregory Crane) SESSION: Searching across language, time, and space Harvesting translingual vocabulary mappings for multilingual digital libraries (Ray R. Larson, Fredric Gey, Aitao Chen) - Detecting events with date and place information in unstructured text (David A. Smith) - Using sharable ontology to retrieve historical images (Von-Wun Soo, Chen-Yu Lee, Jaw Jium Yeh, Ching-chih Chen) - Towards an electronic variorum edition of Cervantes' Don Quixote:: visualizations that support preparation (Rajiv Kochumman, Carlos Monroy, Richard Furuta, Arpita Goenka, Eduardo Urbina, Erendira Melgoza)
    SESSION: NSDL Core services in the architecture of the national science digital library (NSDL) (Carl Lagoze, William Arms, Stoney Gan, Diane Hillmann, Christopher Ingram, Dean Krafft, Richard Marisa, Jon Phipps, John Saylor, Carol Terrizzi, Walter Hoehn, David Millman, James Allan, Sergio Guzman-Lara, Tom Kalt) - Creating virtual collections in digital libraries: benefits and implementation issues (Gary Geisler, Sarah Giersch, David McArthur, Marty McClelland) - Ontology services for curriculum development in NSDL (Amarnath Gupta, Bertram Ludäscher, Reagan W. Moore) - Interactive digital library resource information system: a web portal for digital library education (Ahmad Rafee Che Kassim, Thomas R. Kochtanek) SESSION: Digital library communities and change Cross-cultural usability of the library metaphor (Elke Duncker) - Trust and epistemic communities in biodiversity data sharing (Nancy A. Van House) - Evaluation of digital community information systems (K. T. Unruh, K. E. Pettigrew, J. C. Durrance) - Adapting digital libraries to continual evolution (Bruce R. Barkstrom, Melinda Finch, Michelle Ferebee, Calvin Mackey) SESSION: Models and tools for generating digital libraries Localizing experience of digital content via structural metadata (Naomi Dushay) - Collection synthesis (Donna Bergmark) - 5SL: a language for declarative specification and generation of digital libraries (Marcos André, Gonçalves, Edward A. Fox) SESSION: Novel user interfaces A digital library of conversational expressions: helping profoundly disabled users communicate (Hayley Dunlop, Sally Jo Cunningham, Matt Jones) - Enhancing the ENVISION interface for digital libraries (Jun Wang, Abhishek Agrawal, Anil Bazaza, Supriya Angle, Edward A. Fox, Chris North) - A wearable digital library of personal conversations (Wei-hao Lin, Alexander G. Hauptmann) - Collaborative visual interfaces to digital libraries (Katy Börner, Ying Feng, Tamara McMahon) - Binding browsing and reading activities in a 3D digital library (Pierre Cubaud, Pascal Stokowski, Alexandre Topol)
    SESSION: Federating and harvesting metadata DP9: an OAI gateway service for web crawlers (Xiaoming Liu, Kurt Maly, Mohammad Zubair, Michael L. Nelson) - The Greenstone plugin architecture (Ian H. Witten, David Bainbridge, Gordon Paynter, Stefan Boddie) - Building FLOW: federating libraries on the web (Anna Keller Gold, Karen S. Baker, Jean-Yves LeMeur, Kim Baldridge) - JAFER ToolKit project: interfacing Z39.50 and XML (Antony Corfield, Matthew Dovey, Richard Mawby, Colin Tatham) - Schema extraction from XML collections (Boris Chidlovskii) - Mirroring an OAI archive on the I2-DSI channel (Ashwini Pande, Malini Kothapalli, Ryan Richardson, Edward A. Fox) SESSION: Music digital libraries HMM-based musical query retrieval (Jonah Shifrin, Bryan Pardo, Colin Meek, William Birmingham) - A comparison of melodic database retrieval techniques using sung queries (Ning Hu, Roger B. Dannenberg) - Enhancing access to the levy sheet music collection: reconstructing full-text lyrics from syllables (Brian Wingenroth, Mark Patton, Tim DiLauro) - Evaluating automatic melody segmentation aimed at music information retrieval (Massimo Melucci, Nicola Orio) SESSION: Preserving, securing, and assessing digital libraries A methodology and system for preserving digital data (Raymond A. Lorie) - Modeling web data (James C. French) - An evaluation model for a digital library services tool (Jim Dorward, Derek Reinke, Mimi Recker) - Why watermark?: the copyright need for an engineering solution (Michael Seadle, J. R. Deller, Jr., Aparna Gurijala) SESSION: Image and cultural digital libraries Time as essence for photo browsing through personal digital libraries (Adrian Graham, Hector Garcia-Molina, Andreas Paepcke, Terry Winograd) - Toward a distributed terabyte text retrieval system in China-US million book digital library (Bin Liu, Wen Gao, Ling Zhang, Tie-jun Huang, Xiao-ming Zhang, Jun Cheng) - Enhanced perspectives for historical and cultural documentaries using informedia technologies (Howard D. Wactlar, Ching-chih Chen) - Interfaces for palmtop image search (Mark Derthick)
    SESSION: Digital libraries for spatial data The ADEPT digital library architecture (Greg Janée, James Frew) - G-Portal: a map-based digital library for distributed geospatial and georeferenced resources (Ee-Peng Lim, Dion Hoe-Lian Goh, Zehua Liu, Wee-Keong Ng, Christopher Soo-Guan Khoo, Susan Ellen Higgins) PANEL SESSION: Panels You mean I have to do what with whom: statewide museum/library DIGI collaborative digitization projects---the experiences of California, Colorado & North Carolina (Nancy Allen, Liz Bishoff, Robin Chandler, Kevin Cherry) - Overcoming impediments to effective health and biomedical digital libraries (William Hersh, Jan Velterop, Alexa McCray, Gunther Eynsenbach, Mark Boguski) - The challenges of statistical digital libraries (Cathryn Dippo, Patricia Cruse, Ann Green, Carol Hert) - Biodiversity and biocomplexity informatics: policy and implementation science versus citizen science (P. Bryan Heidorn) - Panel on digital preservation (Joyce Ray, Robin Dale, Reagan Moore, Vicky Reich, William Underwood, Alexa T. McCray) - NSDL: from prototype to production to transformational national resource (William Y. Arms, Edward Fox, Jeanne Narum, Ellen Hoffman) - How important is metadata? (Hector Garcia-Molina, Diane Hillmann, Carl Lagoze, Elizabeth Liddy, Stuart Weibel) - Planning for future digital libraries programs (Stephen M. Griffin) DEMONSTRATION SESSION: Demonstrations u.a.: FACET: thesaurus retrieval with semantic term expansion (Douglas Tudhope, Ceri Binding, Dorothee Blocks, Daniel Cunliffe) - MedTextus: an intelligent web-based medical meta-search system (Bin Zhu, Gondy Leroy, Hsinchun Chen, Yongchi Chen) POSTER SESSION: Posters TUTORIAL SESSION: Tutorials u.a.: Thesauri and ontologies in digital libraries: 1. structure and use in knowledge-based assistance to users (Dagobert Soergel) - How to build a digital library using open-source software (Ian H. Witten) - Thesauri and ontologies in digital libraries: 2. design, evaluation, and development (Dagobert Soergel) WORKSHOP SESSION: Workshops Document search interface design for large-scale collections and intelligent access (Javed Mostafa) - Visual interfaces to digital libraries (Katy Börner, Chaomei Chen) - Text retrieval conference (TREC) genomics pre-track workshop (William Hersh)
  17. Research and advanced technology for digital libraries : 11th European conference, ECDL 2007 / Budapest, Hungary, September 16-21, 2007, proceedings (2007) 0.01
    0.0059951874 = product of:
      0.02398075 = sum of:
        0.02398075 = product of:
          0.0479615 = sum of:
            0.0479615 = weight(_text_:processing in 2430) [ClassicSimilarity], result of:
              0.0479615 = score(doc=2430,freq=4.0), product of:
                0.18956426 = queryWeight, product of:
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.046827413 = queryNorm
                0.2530092 = fieldWeight in 2430, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2430)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    LCSH
    Document Preparation and Text Processing
    Subject
    Document Preparation and Text Processing
  18. Information und Wissen : global, sozial und frei? Proceedings des 12. Internationalen Symposiums für Informationswissenschaft (ISI 2011) ; Hildesheim, 9. - 11. März 2011 (2010) 0.01
    0.005599941 = product of:
      0.022399765 = sum of:
        0.022399765 = weight(_text_:data in 5190) [ClassicSimilarity], result of:
          0.022399765 = score(doc=5190,freq=6.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.15127754 = fieldWeight in 5190, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.01953125 = fieldNorm(doc=5190)
      0.25 = coord(1/4)
    
    Content
    - Infometrics & Representations Steffen Hennicke, Marlies Olensky, Viktor de Boer, Antoine Isaac, Jan Wielemaker: A data model for cross-domain data representation Stefanie Haustein: Wissenschaftliche Zeitschriften im Web 2.0 " Philipp Leinenkugel, Werner Dees, Marc Rittberger: Abdeckung erziehungswissenschaftlicher Zeitschriften in Google Scholar - Information Retrieval Ari Pirkola: Constructing Topic-specific Search Keyphrase: Suggestion Tools for Web Information Retrieval Philipp Mayr, Peter Mutschke, Vivien Petras, Philipp Schaer, York Sure: Applying Science Models for Search Daniela Becks, Thomas Mandl, Christa Womser-Hacker: Spezielle Anforderungen bei der Evaluierung von Patent-Retrieval-Systemen Andrea Ernst-Gerlach, Dennis Korbar, Ära Awakian: Entwicklung einer Benutzeroberfläche zur interaktiven Regelgenerierung für die Suche in historischen Dokumenten - Multimedia Peter Schultes, Franz Lehner, Harald Kosch: Effects of real, media and presentation time in annotated video Marc Ritter, Maximilian Eibl: Ein erweiterbares Tool zur Annotation von Videos Margret Plank: AV-Portal für wissenschaftliche Filme: Analyse der Nutzerbedarfe Achim Oßwald: Significant properties digitaler Objekte
    - Information Professionals & Usage Rahmatollah Fattahi, Mohaddeseh Dokhtesmati, Maryam Saberi: A survey of internet searching skills among intermediate school students: How librarians can help Matthias Görtz: Kontextspezifische Erhebung von aufgabenbezogenem Informationssuchverhalten Jürgen Reischer, Daniel Lottes, Florian Meier, Matthias Stirner: Evaluation von Summarizing-Systemen Robert Mayo Hayes, Karin Karlics, Christian Schlögl: Bedarf an Informationsspezialisten in wissensintensiven Branchen der österreichischen Volkswirtschaft - User Experience fit Behavior Isto Huvila: Mining qualitative data on human information behaviour from the Web Rahel Birri Blezon, Rene Schneider: The Social Persona Approach Elena Shpilka, Ralph Koelle, Wolfgang Semar: "Mobile Tagging": Konzeption und Implementierung eines mobilen Informationssystems mit 2D-Tags Johannes Baeck, Sabine Wiem, Ralph Kölle, Thomas Mandl: User Interface Prototyping Nadine Mahrholz, Thomas Mandl, Joachim Griesbaum: Analyse und Evaluierung der Nutzung von Sitelinks Bernard Bekavac, Sonja Öttl, Thomas Weinhold: Online-Beratungskomponente für die Auswahl von Usability-Evaluationsmethoden
  19. Stuckenschmidt, H.; Harmelen, F. van: Information sharing on the semantic web (2005) 0.01
    0.005299047 = product of:
      0.021196188 = sum of:
        0.021196188 = product of:
          0.042392377 = sum of:
            0.042392377 = weight(_text_:processing in 2789) [ClassicSimilarity], result of:
              0.042392377 = score(doc=2789,freq=2.0), product of:
                0.18956426 = queryWeight, product of:
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.046827413 = queryNorm
                0.22363065 = fieldWeight in 2789, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.048147 = idf(docFreq=2097, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2789)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Series
    Advanced information and knowledge processing
  20. Semantic digital libraries (2009) 0.01
    0.0051730038 = product of:
      0.020692015 = sum of:
        0.020692015 = weight(_text_:data in 3371) [ClassicSimilarity], result of:
          0.020692015 = score(doc=3371,freq=2.0), product of:
            0.14807065 = queryWeight, product of:
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.046827413 = queryNorm
            0.1397442 = fieldWeight in 3371, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1620505 = idf(docFreq=5088, maxDocs=44218)
              0.03125 = fieldNorm(doc=3371)
      0.25 = coord(1/4)
    
    Abstract
    Libraries have always been an inspiration for the standards and technologies developed by semantic web activities. However, except for the Dublin Core specification, semantic web and social networking technologies have not been widely adopted and further developed by major digital library initiatives and projects. Yet semantic technologies offer a new level of flexibility, interoperability, and relationships for digital repositories. Kruk and McDaniel present semantic web-related aspects of current digital library activities, and introduce their functionality; they show examples ranging from general architectural descriptions to detailed usages of specific ontologies, and thus stimulate the awareness of researchers, engineers, and potential users of those technologies. Their presentation is completed by chapters on existing prototype systems such as JeromeDL, BRICKS, and Greenstone, as well as a look into the possible future of semantic digital libraries. This book is aimed at researchers and graduate students in areas like digital libraries, the semantic web, social networks, and information retrieval. This audience will benefit from detailed descriptions of both today's possibilities and also the shortcomings of applying semantic web technologies to large digital repositories of often unstructured data.

Years

Languages

  • e 21
  • d 8

Types

  • m 30
  • s 13
  • d 1
  • More… Less…

Subjects

Classifications