Search (3722 results, page 1 of 187)

  1. Alexander, M.: Digitising books, manuscripts and scholarly materials : preparation, handling, scanning, recognition, compression, storage formats (1998) 0.20
    0.20035386 = product of:
      0.40070772 = sum of:
        0.40070772 = sum of:
          0.34482676 = weight(_text_:compression in 3686) [ClassicSimilarity], result of:
            0.34482676 = score(doc=3686,freq=4.0), product of:
              0.3771246 = queryWeight, product of:
                7.314861 = idf(docFreq=79, maxDocs=44218)
                0.05155595 = queryNorm
              0.9143576 = fieldWeight in 3686, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                7.314861 = idf(docFreq=79, maxDocs=44218)
                0.0625 = fieldNorm(doc=3686)
          0.055880956 = weight(_text_:22 in 3686) [ClassicSimilarity], result of:
            0.055880956 = score(doc=3686,freq=2.0), product of:
              0.18054016 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05155595 = queryNorm
              0.30952093 = fieldWeight in 3686, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=3686)
      0.5 = coord(1/2)
    
    Abstract
    The British Library's Initiatives for Access programme (1993-) aims to identify the impact and value of digital and networking technologies on the Library's collections and services. Describes the projects: the Electronic Beowulf, digitisation of ageing microfilm, digital photographic images, and use of the Excalibur retrieval software. Examines the ways in which the issues of preparation, scanning, and storage have been tackled, and problems raised by use of recognition technologies and compression
    Date
    22. 5.1999 19:00:52
  2. Wolff, J.G.: Computing, cognition and information compression (1993) 0.15
    0.15239334 = product of:
      0.30478668 = sum of:
        0.30478668 = product of:
          0.60957336 = sum of:
            0.60957336 = weight(_text_:compression in 6712) [ClassicSimilarity], result of:
              0.60957336 = score(doc=6712,freq=8.0), product of:
                0.3771246 = queryWeight, product of:
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.05155595 = queryNorm
                1.6163712 = fieldWeight in 6712, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.078125 = fieldNorm(doc=6712)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The storage and processing of information in computers and in brains may often be understood information compression. Reviews what is meant by information and, in particular, what is meant by redundancy, a concept fundamental in all methods for information compression. Describes principles of information compression
  3. Dimitrova, N.; Golshani, F.: Motion recovery for video content classification (1995) 0.15
    0.14985514 = product of:
      0.29971027 = sum of:
        0.29971027 = sum of:
          0.24382932 = weight(_text_:compression in 3834) [ClassicSimilarity], result of:
            0.24382932 = score(doc=3834,freq=2.0), product of:
              0.3771246 = queryWeight, product of:
                7.314861 = idf(docFreq=79, maxDocs=44218)
                0.05155595 = queryNorm
              0.64654845 = fieldWeight in 3834, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                7.314861 = idf(docFreq=79, maxDocs=44218)
                0.0625 = fieldNorm(doc=3834)
          0.055880956 = weight(_text_:22 in 3834) [ClassicSimilarity], result of:
            0.055880956 = score(doc=3834,freq=2.0), product of:
              0.18054016 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05155595 = queryNorm
              0.30952093 = fieldWeight in 3834, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=3834)
      0.5 = coord(1/2)
    
    Abstract
    Discusses the analysis of video for the classification of images in order to develop a video database. Covers compression; motion recovery in digital video; low-level motion extraction; single macroblock tracing; intermediate-level motion analysis; high-level motion analysis; spatiotemporal hierarchical representation; information filtering and digital video; content filtering opertaors; the query language; querying video contents; an architecture for video classification and retrieval; the visual query language VEVA; and implementation of macroblock tracing
    Date
    8. 4.1996 9:22:36
  4. Guenette, D.R.: Document imaging, CD-ROM, and CD-R : a starting point (1996) 0.15
    0.14985514 = product of:
      0.29971027 = sum of:
        0.29971027 = sum of:
          0.24382932 = weight(_text_:compression in 4986) [ClassicSimilarity], result of:
            0.24382932 = score(doc=4986,freq=2.0), product of:
              0.3771246 = queryWeight, product of:
                7.314861 = idf(docFreq=79, maxDocs=44218)
                0.05155595 = queryNorm
              0.64654845 = fieldWeight in 4986, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                7.314861 = idf(docFreq=79, maxDocs=44218)
                0.0625 = fieldNorm(doc=4986)
          0.055880956 = weight(_text_:22 in 4986) [ClassicSimilarity], result of:
            0.055880956 = score(doc=4986,freq=2.0), product of:
              0.18054016 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05155595 = queryNorm
              0.30952093 = fieldWeight in 4986, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=4986)
      0.5 = coord(1/2)
    
    Abstract
    An introduction to technical solutions for the generation and conversion of digital documents, using affordable scanner devices, document imaging systems and OCR technologies with cheap, networkable high storage capacity media such as CD-ROMs and CD-R signals the arrival of CD-ROM based document imaging systems. Describes the processes involved, including: the document imaging process; use of scanners to make bitmaps; data compression; advantages of indexing the images; OCR techniques; and document display. Lists some of the companies providing products and services applicable to CD-ROM and CD-R based document imaging systems
    Date
    6. 9.1996 19:08:22
  5. Cannane, A.; Williams, H.E.: General-purpose compression for efficient retrieval (2001) 0.14
    0.144573 = product of:
      0.289146 = sum of:
        0.289146 = product of:
          0.578292 = sum of:
            0.578292 = weight(_text_:compression in 5705) [ClassicSimilarity], result of:
              0.578292 = score(doc=5705,freq=20.0), product of:
                0.3771246 = queryWeight, product of:
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.05155595 = queryNorm
                1.5334243 = fieldWeight in 5705, product of:
                  4.472136 = tf(freq=20.0), with freq of:
                    20.0 = termFreq=20.0
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5705)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Compression of databases not only reduces space requirements but can also reduce overall retrieval times. In text databases, compression of documents based on semistatic modeling with words has been shown to be both practical and fast. Similarly, for specific applications -such as databases of integers or scientific databases-specially designed semistatic compression schemes work well. We propose a scheme for general-purpose compression that can be applied to all types of data stored in large collections. We describe our approach -which we call RAY-in detail, and show experimentally the compression available, compression and decompression costs, and performance as a stream and random-access technique. We show that, in many cases, RAY achieves better compression than an efficient Huffman scheme and popular adaptive compression techniques, and that it can be used as an efficient general-purpose compression scheme
  6. Alvarado, R.U.: Cataloging Pierre Bourdieu's books (1994) 0.14
    0.1368802 = sum of:
      0.11592485 = product of:
        0.34777454 = sum of:
          0.34777454 = weight(_text_:lcshs in 894) [ClassicSimilarity], result of:
            0.34777454 = score(doc=894,freq=2.0), product of:
              0.52006805 = queryWeight, product of:
                10.087449 = idf(docFreq=4, maxDocs=44218)
                0.05155595 = queryNorm
              0.6687097 = fieldWeight in 894, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                10.087449 = idf(docFreq=4, maxDocs=44218)
                0.046875 = fieldNorm(doc=894)
        0.33333334 = coord(1/3)
      0.020955358 = product of:
        0.041910715 = sum of:
          0.041910715 = weight(_text_:22 in 894) [ClassicSimilarity], result of:
            0.041910715 = score(doc=894,freq=2.0), product of:
              0.18054016 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05155595 = queryNorm
              0.23214069 = fieldWeight in 894, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=894)
        0.5 = coord(1/2)
    
    Abstract
    Subject headings do not always adequately express the subject content of books and other library materials. Whether due to cataloguer error or inadequacy in the authority list, this deficiency makes it difficult for users to access information. In an attempt to solve this problem, the study evaluated the adequacy of the LoC Subject Headings assigned to the 23 books of Pierre Bourdieu, whose philosophical ideas were judged likely to form a good test of the ability of the subject headings to reflect the ideas proposed by the author. The study examined the subject headings given to 22 books, and their translations into English, Spanish, Portuguese, and German, comprising 88 records in OCLC as of Dec 91. It was found that most of the books received headings not corresponding to their content, as the headings were assigned from the functionalist paradigm. In general, LCSHs ignore the conceptual categories of other paradigms, do not match the current vocabulary used by social scientists, and are ideologically biased
  7. Huang, T.; Mehrotra, S.; Ramchandran, K.: Multimedia Access and Retrieval System (MARS) project (1997) 0.13
    0.13112324 = product of:
      0.2622465 = sum of:
        0.2622465 = sum of:
          0.21335067 = weight(_text_:compression in 758) [ClassicSimilarity], result of:
            0.21335067 = score(doc=758,freq=2.0), product of:
              0.3771246 = queryWeight, product of:
                7.314861 = idf(docFreq=79, maxDocs=44218)
                0.05155595 = queryNorm
              0.5657299 = fieldWeight in 758, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                7.314861 = idf(docFreq=79, maxDocs=44218)
                0.0546875 = fieldNorm(doc=758)
          0.048895836 = weight(_text_:22 in 758) [ClassicSimilarity], result of:
            0.048895836 = score(doc=758,freq=2.0), product of:
              0.18054016 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05155595 = queryNorm
              0.2708308 = fieldWeight in 758, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=758)
      0.5 = coord(1/2)
    
    Abstract
    Reports results of the MARS project, conducted at Illinois University, to bring together researchers in the fields of computer vision, compression, information management and database systems with the goal of developing an effective multimedia database management system. Describes the first step, involving the design and implementation of an image retrieval system incorporating novel approaches to image segmentation, representation, browsing and information retrieval supported by the developed system. Points to future directions for the MARS project
    Date
    22. 9.1997 19:16:05
  8. Brandt, R.: Video compression : the why and the how (1993) 0.13
    0.12931003 = product of:
      0.25862005 = sum of:
        0.25862005 = product of:
          0.5172401 = sum of:
            0.5172401 = weight(_text_:compression in 4286) [ClassicSimilarity], result of:
              0.5172401 = score(doc=4286,freq=4.0), product of:
                0.3771246 = queryWeight, product of:
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.05155595 = queryNorm
                1.3715364 = fieldWeight in 4286, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4286)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Describes the technology of video compression as the key to the practical application of CD-ROM storage to multimedia CD-ROM databases
  9. Gates, R.; Bang, S.: Compression and archiving (1993) 0.12
    0.12191466 = product of:
      0.24382932 = sum of:
        0.24382932 = product of:
          0.48765865 = sum of:
            0.48765865 = weight(_text_:compression in 4428) [ClassicSimilarity], result of:
              0.48765865 = score(doc=4428,freq=2.0), product of:
                0.3771246 = queryWeight, product of:
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.05155595 = queryNorm
                1.2930969 = fieldWeight in 4428, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.125 = fieldNorm(doc=4428)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  10. Pennebaker, W.B.; Mitchell, J.L.: JPEG still image data compression standard (1993) 0.12
    0.12191466 = product of:
      0.24382932 = sum of:
        0.24382932 = product of:
          0.48765865 = sum of:
            0.48765865 = weight(_text_:compression in 5251) [ClassicSimilarity], result of:
              0.48765865 = score(doc=5251,freq=2.0), product of:
                0.3771246 = queryWeight, product of:
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.05155595 = queryNorm
                1.2930969 = fieldWeight in 5251, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.125 = fieldNorm(doc=5251)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  11. Jain, A.K.: Image data compression : a review (1981) 0.12
    0.12191466 = product of:
      0.24382932 = sum of:
        0.24382932 = product of:
          0.48765865 = sum of:
            0.48765865 = weight(_text_:compression in 8696) [ClassicSimilarity], result of:
              0.48765865 = score(doc=8696,freq=2.0), product of:
                0.3771246 = queryWeight, product of:
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.05155595 = queryNorm
                1.2930969 = fieldWeight in 8696, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.125 = fieldNorm(doc=8696)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  12. Bookstein, A.; Klein, S.T.: Compression, information theory, and grammars : a unified approach (1990) 0.12
    0.12191466 = product of:
      0.24382932 = sum of:
        0.24382932 = product of:
          0.48765865 = sum of:
            0.48765865 = weight(_text_:compression in 2970) [ClassicSimilarity], result of:
              0.48765865 = score(doc=2970,freq=2.0), product of:
                0.3771246 = queryWeight, product of:
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.05155595 = queryNorm
                1.2930969 = fieldWeight in 2970, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.125 = fieldNorm(doc=2970)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  13. Cheng, K.-S.; Young, G.H.; Wong, K.-F.: ¬A study on word-based and integral-bit Chinese text compression algorithms (1999) 0.12
    0.12191466 = product of:
      0.24382932 = sum of:
        0.24382932 = product of:
          0.48765865 = sum of:
            0.48765865 = weight(_text_:compression in 3056) [ClassicSimilarity], result of:
              0.48765865 = score(doc=3056,freq=8.0), product of:
                0.3771246 = queryWeight, product of:
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.05155595 = queryNorm
                1.2930969 = fieldWeight in 3056, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3056)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Experimental results show that a word-based arithmetic coding scheme can achieve a higher compression performance for Chinese text. However, an arithmetic coding scheme is a fractional-bit compression algorithm which is known to be time comsuming. In this article, we change the direction to study how to cascade the word segmentation model with a faster alternative, the integral-bit compression algorithm. It is shown that the cascaded algorithm is mor suitable for practical usage.
  14. Maguire, P.; Maguire, R.: Consciousness is data compression (2010) 0.12
    0.12191466 = product of:
      0.24382932 = sum of:
        0.24382932 = product of:
          0.48765865 = sum of:
            0.48765865 = weight(_text_:compression in 4972) [ClassicSimilarity], result of:
              0.48765865 = score(doc=4972,freq=8.0), product of:
                0.3771246 = queryWeight, product of:
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.05155595 = queryNorm
                1.2930969 = fieldWeight in 4972, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4972)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    In this article we advance the conjecture that conscious awareness is equivalent to data compression. Algorithmic information theory supports the assertion that all forms of understanding are contingent on compression (Chaitin, 2007). Here, we argue that the experience people refer to as consciousness is the particular form of understanding that the brain provides. We therefore propose that the degree of consciousness of a system can be measured in terms of the amount of data compression it carries out.
  15. Fernandez, C.W.: Semantic relationships between title phrases and LCSH (1991) 0.12
    0.11712619 = product of:
      0.23425238 = sum of:
        0.23425238 = product of:
          0.7027571 = sum of:
            0.7027571 = weight(_text_:lcshs in 509) [ClassicSimilarity], result of:
              0.7027571 = score(doc=509,freq=6.0), product of:
                0.52006805 = queryWeight, product of:
                  10.087449 = idf(docFreq=4, maxDocs=44218)
                  0.05155595 = queryNorm
                1.3512791 = fieldWeight in 509, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  10.087449 = idf(docFreq=4, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=509)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Abstract
    Successful subject searching depends upon the terminology of the search query. This paper compares the terminology of book titles with that of LCSH on a semantic level. Methods of analyzing titles into title phrases that serve as matching units are described. Guidelines used to choose LCSHs with the closest semantic matches to title phrases are set forth. The pairing of title phrases with semantically related LCSHs constitutes the data of the study. Categories representing the types of semantic relationships holding between pairs of terms are defined. Results of the study show that the most common semantic relationship holding between title phrases and LCSH is a hierarchical one where title phrases are more specific than LCSHs
  16. Delfino, E.: ¬The Internet toolkit : file compression and archive utilities (1993) 0.11
    0.10775836 = product of:
      0.21551672 = sum of:
        0.21551672 = product of:
          0.43103343 = sum of:
            0.43103343 = weight(_text_:compression in 6718) [ClassicSimilarity], result of:
              0.43103343 = score(doc=6718,freq=4.0), product of:
                0.3771246 = queryWeight, product of:
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.05155595 = queryNorm
                1.142947 = fieldWeight in 6718, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.078125 = fieldNorm(doc=6718)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    As a result of the combination of high transmission speeds and large data file sizes, many files available over the Internet come in archived and compressed form which need to be decompressed before being read. Discusses the techniques available for file compression and extraction and where to find these utilities on the Internet
  17. Bell, T.C.; Moffat, A.; Nevill-Manning, C.G.; Witten, I.H.; Zobel, J.: Data compression in full-text retrieval system (1993) 0.11
    0.106675334 = product of:
      0.21335067 = sum of:
        0.21335067 = product of:
          0.42670134 = sum of:
            0.42670134 = weight(_text_:compression in 5643) [ClassicSimilarity], result of:
              0.42670134 = score(doc=5643,freq=8.0), product of:
                0.3771246 = queryWeight, product of:
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.05155595 = queryNorm
                1.1314598 = fieldWeight in 5643, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5643)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    When data compression is applied to full-text retrieval systems, intricate relationships emerge between the amount of compression, access speed, and computing resources required. We propose compression methods, and explore corresponding tradeoffs, for all components of static full-text systems such as text databases on CD-ROM. These components include lexical indexes, and the mein text itself. Results are reported on the application of the methods to several substantial full-text databases, and show that a large, unindexed text can be stored, along with indexes that facilitate fast searching, in less than half its original size - at some appreciable cost in primary memory requirements
  18. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.10
    0.10283987 = sum of:
      0.08188452 = product of:
        0.24565354 = sum of:
          0.24565354 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
            0.24565354 = score(doc=562,freq=2.0), product of:
              0.43709192 = queryWeight, product of:
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.05155595 = queryNorm
              0.56201804 = fieldWeight in 562, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                8.478011 = idf(docFreq=24, maxDocs=44218)
                0.046875 = fieldNorm(doc=562)
        0.33333334 = coord(1/3)
      0.020955358 = product of:
        0.041910715 = sum of:
          0.041910715 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
            0.041910715 = score(doc=562,freq=2.0), product of:
              0.18054016 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.05155595 = queryNorm
              0.23214069 = fieldWeight in 562, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=562)
        0.5 = coord(1/2)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  19. Wittan, I.H.; Bell, T.C.; Nevill, C.G.: Indexing and compressing full-text databases for CD-ROM (1991) 0.09
    0.09238355 = product of:
      0.1847671 = sum of:
        0.1847671 = product of:
          0.3695342 = sum of:
            0.3695342 = weight(_text_:compression in 4828) [ClassicSimilarity], result of:
              0.3695342 = score(doc=4828,freq=6.0), product of:
                0.3771246 = queryWeight, product of:
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.05155595 = queryNorm
                0.97987294 = fieldWeight in 4828, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4828)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    CD-ROM is an attractive delivery vehicle for full text databases. Large storage capacity and low access speed, carefully designed indexing structures, including a concordance, are necessary to enable the text to be retrieved efficiently. However, the indexes are sufficiently large that they tax the ability of the main store to hold them when processing queries. The use of compression techniques can substantially increase the volume of text that a disc can accomodate, and substantially decrease the amount of primary storage needed to hold the indexes. Describes a suitable indexing mechanism, and its compression potential using modern compression methods. It is possible to double the amount of text that can be stored on a CD-ROM disc and include a full concordance and indexes as well
  20. Akman, K.I.: ¬A new text compression technique based on natural language structure (1995) 0.09
    0.09238355 = product of:
      0.1847671 = sum of:
        0.1847671 = product of:
          0.3695342 = sum of:
            0.3695342 = weight(_text_:compression in 1860) [ClassicSimilarity], result of:
              0.3695342 = score(doc=1860,freq=6.0), product of:
                0.3771246 = queryWeight, product of:
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.05155595 = queryNorm
                0.97987294 = fieldWeight in 1860, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  7.314861 = idf(docFreq=79, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1860)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Describes a new data compression technique which utilizes some of the common structural characteristics of languages. The proposed algorithm partitions words into their roots and suffixes which are then replaced by shorter bit representations. The method used 3 dictionaries in the from of binary search trees and 1 character array. The first 2 dictionaries are for roots, and the third one is for suffixes. The character array is used for both searching compressible words and coding incompressible words. The number of bits in representing a substring depends on the number of the entries in the dictionary in which the substring is found. The proposed algorithm is implemented in the Turkish language and tested using 3 different text groups with different lenghts. Results indicate a compression factor of up to 47 per cent

Languages

Types

  • a 3112
  • m 352
  • el 163
  • s 142
  • b 40
  • x 35
  • i 23
  • r 17
  • ? 8
  • p 4
  • d 3
  • n 3
  • u 2
  • z 2
  • au 1
  • h 1
  • More… Less…

Themes

Subjects

Classifications