Search (3 results, page 1 of 1)

  • × author_ss:"Borsack, J."
  1. Tagheva, K.; Borsack, J.; Condit, A.: Effects of OCR errors on ranking and feedback using the vector space model (1996) 0.01
    0.0063011474 = product of:
      0.015752869 = sum of:
        0.009437811 = weight(_text_:a in 4951) [ClassicSimilarity], result of:
          0.009437811 = score(doc=4951,freq=6.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.17652355 = fieldWeight in 4951, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=4951)
        0.006315058 = product of:
          0.012630116 = sum of:
            0.012630116 = weight(_text_:information in 4951) [ClassicSimilarity], result of:
              0.012630116 = score(doc=4951,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.1551638 = fieldWeight in 4951, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4951)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Reports on the performance of the vector space model in the presence of optical character recognition (OCR) errors. Average precision and recall is not affected for full text document rankings of the OCR and corrected collections with different weithing combinations. Cosine normalization plays a considerable role in the disparity seen between the collections. Even though feedback improves retrieval for both collections, it can not be used to compensate for OCR errors caused by badly degraded documents
    Source
    Information processing and management. 32(1996) no.3, S.317-327
    Type
    a
  2. Taghva, K.; Borsack, J.; Condit, A.: Evaluation of model-based retrieval effectiveness with OCR text (1996) 0.01
    0.005822873 = product of:
      0.014557183 = sum of:
        0.0067426977 = weight(_text_:a in 4485) [ClassicSimilarity], result of:
          0.0067426977 = score(doc=4485,freq=4.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.12611452 = fieldWeight in 4485, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4485)
        0.007814486 = product of:
          0.015628971 = sum of:
            0.015628971 = weight(_text_:information in 4485) [ClassicSimilarity], result of:
              0.015628971 = score(doc=4485,freq=4.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.1920054 = fieldWeight in 4485, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4485)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Reports on experiments with retrieval from OCR-generated text using systems based on standard models of retrieval. Shows that average precision and recall is not affected by OCR errors across systems for several collections. Both the actual and the simulation experiments include full text and abstract length documents. The ranking and feedback methods associated with the retrieval models are generally not robust enough to deal with OCR errors. OCR errors and garbage strings generated from the mistranslation of graphic objects increase the size of the index significantly. Describes the problems of applying OCR text within an information retrieval environment and offers solutions
    Source
    ACM transactions on information systems. 14(1996) no.1, S.64-93
    Type
    a
  3. Taghva, K.; Borsack, J.; Nartker, T.; Condit, A.: ¬The role of manually-assigned keywords in query expansion (2004) 0.01
    0.005513504 = product of:
      0.01378376 = sum of:
        0.008258085 = weight(_text_:a in 2567) [ClassicSimilarity], result of:
          0.008258085 = score(doc=2567,freq=6.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.1544581 = fieldWeight in 2567, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2567)
        0.005525676 = product of:
          0.011051352 = sum of:
            0.011051352 = weight(_text_:information in 2567) [ClassicSimilarity], result of:
              0.011051352 = score(doc=2567,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.13576832 = fieldWeight in 2567, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2567)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    We report on two types of experiments with respect to manually-assigned keywords to documents in a collection. The first type of experiment examines the usefulness of manually-assigned keywords to automatic feedback. The second type of experiment considers the potential benefits of these keywords to the user as an interactive tool. Several experiments were run and compared. The results of these experiments indicate that there is no gain in average precision when manually-assigned keywords are used for query expansion. Further, manually-assigned keywords did not aid the user as an interactive tool for document understanding.
    Source
    Information processing and management. 40(2004) no.3, S.441-458
    Type
    a