Search (10 results, page 1 of 1)

  • × author_ss:"Blair, D.C."
  1. Blair, D.C.: STAIRS Redux : thoughts on the STAIRS evaluation, ten years after (1996) 0.05
    0.05087889 = product of:
      0.10175778 = sum of:
        0.10175778 = sum of:
          0.0522464 = weight(_text_:retrieval in 3002) [ClassicSimilarity], result of:
            0.0522464 = score(doc=3002,freq=4.0), product of:
              0.15791564 = queryWeight, product of:
                3.024915 = idf(docFreq=5836, maxDocs=44218)
                0.052204985 = queryNorm
              0.33085006 = fieldWeight in 3002, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.024915 = idf(docFreq=5836, maxDocs=44218)
                0.0546875 = fieldNorm(doc=3002)
          0.04951138 = weight(_text_:22 in 3002) [ClassicSimilarity], result of:
            0.04951138 = score(doc=3002,freq=2.0), product of:
              0.18281296 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.052204985 = queryNorm
              0.2708308 = fieldWeight in 3002, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=3002)
      0.5 = coord(1/2)
    
    Abstract
    The test of retrieval effectiveness performed on IBM's STAIRS and reported in 'Communications of the ACM' 10 years ago, continues to be cited frequently in the information retrieval literature. The reasons for the study's continuing pertinence to today's research are discussed, and the political, legal, and commercial aspects of the study are presented. In addition, the method of calculating recall that was used in the STAIRS study is discussed in some detail, especially how it reduces the 5 major types of uncertainty in recall estimations. It is also suggested that this method of recall estimation may serve as the basis for recall estimations that might be truly comparable between systems
    Source
    Journal of the American Society for Information Science. 47(1996) no.1, S.4-22
  2. Blair, D.C.; Maron, M.E.: ¬An evaluation of retrieval effectiveness for a full-text document-retrieval system (1985) 0.03
    0.026388418 = product of:
      0.052776836 = sum of:
        0.052776836 = product of:
          0.10555367 = sum of:
            0.10555367 = weight(_text_:retrieval in 1345) [ClassicSimilarity], result of:
              0.10555367 = score(doc=1345,freq=8.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.6684181 = fieldWeight in 1345, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1345)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Footnote
    Vgl. auch : Salton, G.: Another look ... Comm. ACM 29(1986) S.S.648-656; Blair, D.C.: Full text retrieval ... Int. Class. 13(1986) S.18-23: Blair, D.C., M.E. Maron: Full-text information retrieval ... Inf. proc. man. 26(1990) S.437-447.
  3. Blair, D.C.: ¬An extended relational document retrieval model (1988) 0.02
    0.021110734 = product of:
      0.042221468 = sum of:
        0.042221468 = product of:
          0.084442936 = sum of:
            0.084442936 = weight(_text_:retrieval in 5281) [ClassicSimilarity], result of:
              0.084442936 = score(doc=5281,freq=2.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.5347345 = fieldWeight in 5281, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.125 = fieldNorm(doc=5281)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  4. Blair, D.C.: Language and representation in information retrieval (1991) 0.02
    0.020440381 = product of:
      0.040880762 = sum of:
        0.040880762 = product of:
          0.081761524 = sum of:
            0.081761524 = weight(_text_:retrieval in 1545) [ClassicSimilarity], result of:
              0.081761524 = score(doc=1545,freq=30.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.51775444 = fieldWeight in 1545, product of:
                  5.477226 = tf(freq=30.0), with freq of:
                    30.0 = termFreq=30.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1545)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Information or Document Retrieval is the subject of this book. It is not an introductory book, although it is self-contained in the sense that it is not necessary to have a background in the theory or practice of Information Retrieval in order to understand its arguments. The book presents, as clearly as possible, one particular perspective on Information Retrieval, and attempts to say that certain aspects of the theory or practice of the management of documents are more important than others. The majority of Information Retrieval research has been aimed at the more experimentally tractable small-scale systems, and although much of that work has added greatly to our understanding of Information Retrieval it is becoming increasingly apparent that retrieval systems with large data bases of documents are a fundamentally different genre of systems than small-scale systems. If this is so, which is the thesis of this book, then we must now study large information retrieval systems with the same rigor and intensity that we once studied small-scale systems. Hegel observed that the quantitative growth of any system caused qualitative changes to take place in its structure and processes.
    COMPASS
    Information retrieval
    LCSH
    Information Retrieval
    RSWK
    Information Retrieval
    Subject
    Information retrieval
    Information Retrieval
    Information Retrieval
    Theme
    Verbale Doksprachen im Online-Retrieval
  5. Blair, D.C.: Information retrieval and the philosophy of language (2002) 0.02
    0.019028958 = product of:
      0.038057916 = sum of:
        0.038057916 = product of:
          0.07611583 = sum of:
            0.07611583 = weight(_text_:retrieval in 4283) [ClassicSimilarity], result of:
              0.07611583 = score(doc=4283,freq=26.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.48200315 = fieldWeight in 4283, product of:
                  5.0990195 = tf(freq=26.0), with freq of:
                    26.0 = termFreq=26.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4283)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Information retrieval - the retrieval, primarily, of documents or textual material - is fundamentally a linguistic process. At the very least we must describe what we want and match that description with descriptions of the information that is available to us. Furthermore, when we describe what we want, we must mean something by that description. This is a deceptively simple act, but such linguistic events have been the grist for philosophical analysis since Aristotle. Although there are complexities involved in referring to authors, document types, or other categories of information retrieval context, here I wish to focus an one of the most problematic activities in information retrieval: the description of the intellectual content of information items. And even though I take information retrieval to involve the description and retrieval of written text, what I say here is applicable to any information item whose intellectual content can be described for retrieval-books, documents, images, audio clips, video clips, scientific specimens, engineering schematics, and so forth. For convenience, though, I will refer only to the description and retrieval of documents. The description of intellectual content can go wrong in many obvious ways. We may describe what we want incorrectly; we may describe it correctly but in such general terms that its description is useless for retrieval; or we may describe what we want correctly, but misinterpret the descriptions of available information, and thereby match our description of what we want incorrectly. From a linguistic point of view, we can be misunderstood in the process of retrieval in many ways. Because the philosophy of language deals specifically with how we are understood and mis-understood, it should have some use for understanding the process of description in information retrieval. First, however, let us examine more closely the kinds of misunderstandings that can occur in information retrieval. We use language in searching for information in two principal ways. We use it to describe what we want and to discriminate what we want from other information that is available to us but that we do not want. Description and discrimination together articulate the goals of the information search process; they also delineate the two principal ways in which language can fail us in this process. Van Rijsbergen (1979) was the first to make this distinction, calling them "representation" and "discrimination.""
  6. Blair, D.C.: ¬The challenge of commercial document retrieval : Part II: a strategy for document searching based on identifiable document partitions (2002) 0.02
    0.018282432 = product of:
      0.036564864 = sum of:
        0.036564864 = product of:
          0.07312973 = sum of:
            0.07312973 = weight(_text_:retrieval in 2576) [ClassicSimilarity], result of:
              0.07312973 = score(doc=2576,freq=6.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.46309367 = fieldWeight in 2576, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2576)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    In the first article of this three-part series, three factors were identified as major influences on the effectiveness of document retrieval, where document retrieval is concerned with the access to intellectual content. These factors are, search exhaustivity, determinacy of representation and document collection size. In this article, some of the implications of these factors for document searching are discussed, and a two-part searching procedure based on identifiable partitions is proposed.
  7. Blair, D.C.: Full text retrieval : Evaluation and implications (1986) 0.02
    0.01770189 = product of:
      0.03540378 = sum of:
        0.03540378 = product of:
          0.07080756 = sum of:
            0.07080756 = weight(_text_:retrieval in 2047) [ClassicSimilarity], result of:
              0.07080756 = score(doc=2047,freq=10.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.44838852 = fieldWeight in 2047, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2047)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Recently, a detailed evaluation of a large, operational full-text document retrieval system was reported in the literature. Values of precision and recall were estimated usind traditional statistical sampling methods and blind evaluation procedures. The results of this evaluation demonstrated that the system tested was retrieving less then 20% of the relevant documents when the searchers believed it was retrieving over 75% of the relevant documents. This evaluation is described including some data not reported in the original article. Also discussed are the implications which this study has for how the subjects of documents should be represented, as well as the importance of rigorous retrieval evaluations for the furtherhance of information retrieval research
    Footnote
    Vgl.: Blair, D.C., M.E. Maron: An evaluation ... Comm. ACM 28(1985) S.280-299; Salton, G.: Another look ... Comm. ACM 29(1986) S.648-656; Blair, D.C., M.E. Maron: Full-text information retrieval ... Inf. Proc. Man. 26(1990) S.437-447.
  8. Blair, D.C.: ¬The challenge of commercial document retrieval : Part I: Major issues, and a framework based on search exhaustivity, determinacy of representation and document collection size (2002) 0.02
    0.017504096 = product of:
      0.035008192 = sum of:
        0.035008192 = product of:
          0.070016384 = sum of:
            0.070016384 = weight(_text_:retrieval in 2580) [ClassicSimilarity], result of:
              0.070016384 = score(doc=2580,freq=22.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.44337842 = fieldWeight in 2580, product of:
                  4.690416 = tf(freq=22.0), with freq of:
                    22.0 = termFreq=22.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2580)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    With the growing focus on what is collectively known as "knowledge management", a shift continues to take place in commercial information system development: a shift away from the well-understood data retrieval/database model, to the more complex and challenging development of commercial document/information retrieval models. While document retrieval has had a long and rich legacy of research, its impact on commercial applications has been modest. At the enterprise level most large organizations have little understanding of, or commitment to, high quality document access and management. Part of the reason for this is that we still do not have a good framework for understanding the major factors which affect the performance of large-scale corporate document retrieval systems. The thesis of this discussion is that document retrieval - specifically, access to intellectual content - is a complex process which is most strongly influenced by three factors: the size of the document collection; the type of search (exhaustive, existence or sample); and, the determinacy of document representation. Collectively, these factors can be used to provide a useful framework for, or taxonomy of, document retrieval, and highlight some of the fundamental issues facing the design and development of commercial document retrieval systems. This is the first of a series of three articles. Part II (D.C. Blair, The challenge of commercial document retrieval. Part II. A strategy for document searching based on identifiable document partitions, Information Processing and Management, 2001b, this issue) will discuss the implications of this framework for search strategy, and Part III (D.C. Blair, Some thoughts on the reported results of Text REtrieval Conference (TREC), Information Processing and Management, 2002, forthcoming) will consider the importance of the TREC results for our understanding of operating information retrieval systems.
  9. Blair, D.C.; Maron, M.E.: Full-text information retrieval : further analysis and clarification (1990) 0.01
    0.014927543 = product of:
      0.029855086 = sum of:
        0.029855086 = product of:
          0.05971017 = sum of:
            0.05971017 = weight(_text_:retrieval in 2046) [ClassicSimilarity], result of:
              0.05971017 = score(doc=2046,freq=4.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.37811437 = fieldWeight in 2046, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2046)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    In 1985, an article by Blair and Maron described a detailed evaluation of the effectiveness of an operational full text retrieval system used to support the defense of a large corporate lawsuit. The following year Salton published an article which called into question the conclusions of the 1985 study. The following article briefly reviews the initial study, replies to the objections raised by the secon article, and clarifies several confusions and misunderstandings of the 1985 study
  10. Blair, D.C.: Some thoughts on the reported results of TREC (2002) 0.01
    0.014927543 = product of:
      0.029855086 = sum of:
        0.029855086 = product of:
          0.05971017 = sum of:
            0.05971017 = weight(_text_:retrieval in 2583) [ClassicSimilarity], result of:
              0.05971017 = score(doc=2583,freq=4.0), product of:
                0.15791564 = queryWeight, product of:
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.052204985 = queryNorm
                0.37811437 = fieldWeight in 2583, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.024915 = idf(docFreq=5836, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2583)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The periodic TRECs - Text REtrieval Conferences - have reported the results of a variety of recall studies in large-scale document retrieval. While the efforts of TREC are noteworthy and laudable, there are reasons why its results, especially the recall values which are central to its conclusions, should be accepted with some caution.