Search (59 results, page 1 of 3)

  • × theme_ss:"Automatisches Indexieren"
  1. Hodges, P.R.: Keyword in title indexes : effectiveness of retrieval in computer searches (1983) 0.03
    0.031028897 = product of:
      0.09308669 = sum of:
        0.07653132 = weight(_text_:searching in 5001) [ClassicSimilarity], result of:
          0.07653132 = score(doc=5001,freq=6.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.541893 = fieldWeight in 5001, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5001)
        0.016555373 = product of:
          0.033110745 = sum of:
            0.033110745 = weight(_text_:22 in 5001) [ClassicSimilarity], result of:
              0.033110745 = score(doc=5001,freq=2.0), product of:
                0.1222562 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03491209 = queryNorm
                0.2708308 = fieldWeight in 5001, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5001)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    A study was done to test the effectiveness of retrieval using title word searching. It was based on actual search profiles used in the Mechanized Information Center at Ohio State University, in order ro replicate as closely as possible actual searching conditions. Fewer than 50% of the relevant titles were retrieved by keywords in titles. The low rate of retrieval can be attributes to three sources: titles themselves, user and information specialist ignorance of the subject vocabulary in use, and to general language problems. Across fields it was found that the social sciences had the best retrieval rate, with science having the next best, and arts and humanities the lowest. Ways to enhance and supplement keyword in title searching on the computer and in printed indexes are discussed.
    Date
    14. 3.1996 13:22:21
  2. Humphrey, S.M.: Automatic indexing of documents from journal descriptors : a preliminary investigation (1999) 0.02
    0.023941021 = product of:
      0.07182306 = sum of:
        0.03787318 = weight(_text_:searching in 3769) [ClassicSimilarity], result of:
          0.03787318 = score(doc=3769,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.26816747 = fieldWeight in 3769, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.046875 = fieldNorm(doc=3769)
        0.03394988 = product of:
          0.06789976 = sum of:
            0.06789976 = weight(_text_:etc in 3769) [ClassicSimilarity], result of:
              0.06789976 = score(doc=3769,freq=2.0), product of:
                0.18910104 = queryWeight, product of:
                  5.4164915 = idf(docFreq=533, maxDocs=44218)
                  0.03491209 = queryNorm
                0.35906604 = fieldWeight in 3769, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.4164915 = idf(docFreq=533, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3769)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    A new, fully automated approach for indedexing documents is presented based on associating textwords in a training set of bibliographic citations with the indexing of journals. This journal-level indexing is in the form of a consistent, timely set of journal descriptors (JDs) indexing the individual journals themselves. This indexing is maintained in journal records in a serials authority database. The advantage of this novel approach is that the training set does not depend on previous manual indexing of thousands of documents (i.e., any such indexing already in the training set is not used), but rather the relatively small intellectual effort of indexing at the journal level, usually a matter of a few thousand unique journals for which retrospective indexing to maintain consistency and currency may be feasible. If successful, JD indexing would provide topical categorization of documents outside the training set, i.e., journal articles, monographs, Web documents, reports from the grey literature, etc., and therefore be applied in searching. Because JDs are quite general, corresponding to subject domains, their most problable use would be for improving or refining search results
  3. Luhn, H.P.: ¬A statistical approach to the mechanical encoding and searching of literary information (1957) 0.02
    0.016832525 = product of:
      0.100995146 = sum of:
        0.100995146 = weight(_text_:searching in 5453) [ClassicSimilarity], result of:
          0.100995146 = score(doc=5453,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.7151132 = fieldWeight in 5453, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.125 = fieldNorm(doc=5453)
      0.16666667 = coord(1/6)
    
  4. Garfield, E.: KeyWords Plus : ISI's breakthrough retrieval method (1990) 0.01
    0.0147284595 = product of:
      0.088370755 = sum of:
        0.088370755 = weight(_text_:searching in 4345) [ClassicSimilarity], result of:
          0.088370755 = score(doc=4345,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.6257241 = fieldWeight in 4345, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.109375 = fieldNorm(doc=4345)
      0.16666667 = coord(1/6)
    
    Issue
    Pt.1: Expanding your searching power on Current Contents on Diskette.
  5. Pritchard, J.: Information retrieval : smarter indexing (1991) 0.01
    0.010520328 = product of:
      0.06312197 = sum of:
        0.06312197 = weight(_text_:searching in 4890) [ClassicSimilarity], result of:
          0.06312197 = score(doc=4890,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.44694576 = fieldWeight in 4890, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.078125 = fieldNorm(doc=4890)
      0.16666667 = coord(1/6)
    
    Abstract
    Describes full text retrieval (FTR) which indexes every occurrence of every word except defined 'stop' words. This permits much more sophisticated searching than with keyword indexing. Also discusses document imaging processing (DIP). Lists suppliers and users of the software and describes the experiences of ESOO's Planning Division with Computer Intertrade Ltd. (CIL) ImagePro DIP and their operational practices
  6. Samstag-Schnock, U.; Meadow, C.T.: PBS: an ecomical natural language query interpreter (1993) 0.01
    0.008416262 = product of:
      0.050497573 = sum of:
        0.050497573 = weight(_text_:searching in 5091) [ClassicSimilarity], result of:
          0.050497573 = score(doc=5091,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.3575566 = fieldWeight in 5091, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.0625 = fieldNorm(doc=5091)
      0.16666667 = coord(1/6)
    
    Abstract
    Reports on the design and implementation of the information searching and retrieval software, PBS (Parsing, Boolean recognition, Stemming) for the front end OAK 2, a new version of OAK developed at Toronto Univ. OAK 2 is a research tool for user behaviour studies. PBS receives natural language search statements from an end user and identifies search facets and implied Boolean logic operators
  7. Can, F.: Incremental clustering for dynamic information processing (1993) 0.01
    0.008416262 = product of:
      0.050497573 = sum of:
        0.050497573 = weight(_text_:searching in 6627) [ClassicSimilarity], result of:
          0.050497573 = score(doc=6627,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.3575566 = fieldWeight in 6627, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.0625 = fieldNorm(doc=6627)
      0.16666667 = coord(1/6)
    
    Abstract
    Clustering of very large document databases is useful for both searching and browsing. The periodic updating of clusters is required due to the dynamic nature of databases. Introduces an algorithm for incremental clustering and discusses the complexity and cost of analysis of the algorithm together with an investigation of its expected behaviour. Shows through empirical testing that the algortihm achieves cost effectiveness and generates statistically valid clusters that are compatible with those of reclustering. The experimental evidence shows that the algorithm creates an effective and effecient retrieval environment
  8. Polity, Y.: Vers une ergonomie linguistique (1994) 0.01
    0.008416262 = product of:
      0.050497573 = sum of:
        0.050497573 = weight(_text_:searching in 36) [ClassicSimilarity], result of:
          0.050497573 = score(doc=36,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.3575566 = fieldWeight in 36, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.0625 = fieldNorm(doc=36)
      0.16666667 = coord(1/6)
    
    Abstract
    Analyzed a special type of man-mchine interaction, that of searching an information system with natural language. A model for full text processing for information retrieval was proposed that considered the system's users and how they employ information. Describes how INIST (the National Institute for Scientific and Technical Information) is developing computer assisted indexing as an aid to improving relevance when retrieving information from bibliographic data banks
  9. Alexander, M.: Automatic indexing of document images using Excalibur EFS (1995) 0.01
    0.008416262 = product of:
      0.050497573 = sum of:
        0.050497573 = weight(_text_:searching in 1911) [ClassicSimilarity], result of:
          0.050497573 = score(doc=1911,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.3575566 = fieldWeight in 1911, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.0625 = fieldNorm(doc=1911)
      0.16666667 = coord(1/6)
    
    Abstract
    Discusses research into the application of adaptive pattern recognition technology to enable effective retrieval from scanned document images. Describes application at the British Library of Excalibur EFS software which uses adaptive pattern recognition technology to provide access to digital information in its native forms, fuzzy searching retrieval and automatic indexing capabilities. It was used to make specialist printed catalogues and indexes accessible on computer via content based indexes
  10. Alexander, M.: Retrieving digital data with fuzzy matching (1996) 0.01
    0.008416262 = product of:
      0.050497573 = sum of:
        0.050497573 = weight(_text_:searching in 6961) [ClassicSimilarity], result of:
          0.050497573 = score(doc=6961,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.3575566 = fieldWeight in 6961, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.0625 = fieldNorm(doc=6961)
      0.16666667 = coord(1/6)
    
    Abstract
    Briefly describes the Excalibur EFS system which makes use of adaptive pattern recognition technology as an aid to automatic indexing and how it is being tested at the British Library for the indexing and retrieval of scanned images from the library's holdings. Notes how Excalibur EFS can support a wide degree of fuzzy searching, compensate for the errors produced by OCR conversion of scanned images, reduce the costs of indexing, and require far less storage space than more traditional indexes
  11. Stegentritt, E.: Evaluationsresultate des mehrsprachigen Suchsystems CANAL/LS (1998) 0.01
    0.008416262 = product of:
      0.050497573 = sum of:
        0.050497573 = weight(_text_:searching in 7216) [ClassicSimilarity], result of:
          0.050497573 = score(doc=7216,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.3575566 = fieldWeight in 7216, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.0625 = fieldNorm(doc=7216)
      0.16666667 = coord(1/6)
    
    Abstract
    The search system CANAL/LS simplifies the searching of library catalogues by analyzing search questions linguistically and translating them if required. The linguistic analysis reduces the search question words to their basic forms so that they can be compared with basic title forms. Consequently all variants of words and parts of compounds in German can be found. Presents the results of an analysis of search questions in a catalogue of 45.000 titles in the field of psychology
  12. Hirawa, M.: Role of keywords in the network searching era (1998) 0.01
    0.008416262 = product of:
      0.050497573 = sum of:
        0.050497573 = weight(_text_:searching in 3446) [ClassicSimilarity], result of:
          0.050497573 = score(doc=3446,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.3575566 = fieldWeight in 3446, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.0625 = fieldNorm(doc=3446)
      0.16666667 = coord(1/6)
    
  13. Pfeifer, U.; Fuhr, N.; Huynh, T.: Searching structured documents with the enhanced retrieval functionality of freeWAIS-sf and SFgate (1995) 0.01
    0.0073642298 = product of:
      0.044185378 = sum of:
        0.044185378 = weight(_text_:searching in 2214) [ClassicSimilarity], result of:
          0.044185378 = score(doc=2214,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.31286204 = fieldWeight in 2214, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2214)
      0.16666667 = coord(1/6)
    
  14. Shafer, K.: Scorpion Project explores using Dewey to organize the Web (1996) 0.01
    0.0073642298 = product of:
      0.044185378 = sum of:
        0.044185378 = weight(_text_:searching in 6750) [ClassicSimilarity], result of:
          0.044185378 = score(doc=6750,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.31286204 = fieldWeight in 6750, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.0546875 = fieldNorm(doc=6750)
      0.16666667 = coord(1/6)
    
    Abstract
    As the amount of accessible information on the WWW increases, so will the cost of accessing it, even if search servcies remain free, due to the increasing amount of time users will have to spend to find needed items. Considers what the seemingly unorganized Web and the organized world of libraries can offer each other. The OCLC Scorpion Project is attempting to combine indexing and cataloguing, specifically focusing on building tools for automatic subject recognition using the technqiues of library science and information retrieval. If subject headings or concept domains can be automatically assigned to electronic items, improved filtering tools for searching can be produced
  15. SIGIR'92 : Proceedings of the 15th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (1992) 0.01
    0.0063776104 = product of:
      0.03826566 = sum of:
        0.03826566 = weight(_text_:searching in 6671) [ClassicSimilarity], result of:
          0.03826566 = score(doc=6671,freq=6.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.2709465 = fieldWeight in 6671, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.02734375 = fieldNorm(doc=6671)
      0.16666667 = coord(1/6)
    
    Content
    HARMAN, D.: Relevance feedback revisited; AALBERSBERG, I.J.: Incremental relevance feedback; TAGUE-SUTCLIFFE, J.: Measuring the informativeness of a retrieval process; LEWIS, D.D.: An evaluation of phrasal and clustered representations on a text categorization task; BLOSSEVILLE, M.J., G. HÉBRAIL, M.G. MONTEIL u. N. PÉNOT: Automatic document classification: natural language processing, statistical analysis, and expert system techniques used together; MASAND, B., G. LINOFF u. D. WALTZ: Classifying news stories using memory based reasoning; KEEN, E.M.: Term position ranking: some new test results; CROUCH, C.J. u. B. YANG: Experiments in automatic statistical thesaurus construction; GREFENSTETTE, G.: Use of syntactic context to produce term association lists for text retrieval; ANICK, P.G. u. R.A. FLYNN: Versioning of full-text information retrieval system; BURKOWSKI, F.J.: Retrieval activities in a database consisting of heterogeneous collections; DEERWESTER, S.C., K. WACLENA u. M. LaMAR: A textual object management system; NIE, J.-Y.:Towards a probabilistic modal logic for semantic-based information retrieval; WANG, A.W., S.K.M. WONG u. Y.Y. YAO: An analysis of vector space models based on computational geometry; BARTELL, B.T., G.W. COTTRELL u. R.K. BELEW: Latent semantic indexing is an optimal special case of multidimensional scaling; GLAVITSCH, U. u. P. SCHÄUBLE: A system for retrieving speech documents; MARGULIS, E.L.: N-Poisson document modelling; HESS, M.: An incrementally extensible document retrieval system based on linguistics and logical principles; COOPER, W.S., F.C. GEY u. D.P. DABNEY: Probabilistic retrieval based on staged logistic regression; FUHR, N.: Integration of probabilistic fact and text retrieval; CROFT, B., L.A. SMITH u. H. TURTLE: A loosely-coupled integration of a text retrieval system and an object-oriented database system; DUMAIS, S.T. u. J. NIELSEN: Automating the assignement of submitted manuscripts to reviewers; GOST, M.A. u. M. MASOTTI: Design of an OPAC database to permit different subject searching accesses; ROBERTSON, A.M. u. P. WILLETT: Searching for historical word forms in a database of 17th century English text using spelling correction methods; FAX, E.A., Q.F. CHEN u. L.S. HEATH: A faster algorithm for constructing minimal perfect hash functions; MOFFAT, A. u. J. ZOBEL: Parameterised compression for sparse bitmaps; GRANDI, F., P. TIBERIO u. P. Zezula: Frame-sliced patitioned parallel signature files; ALLEN, B.: Cognitive differences in end user searching of a CD-ROM index; SONNENWALD, D.H.: Developing a theory to guide the process of designing information retrieval systems; CUTTING, D.R., J.O. PEDERSEN, D. KARGER, u. J.W. TUKEY: Scatter/ Gather: a cluster-based approach to browsing large document collections; CHALMERS, M. u. P. CHITSON: Bead: Explorations in information visualization; WILLIAMSON, C. u. B. SHNEIDERMAN: The dynamic HomeFinder: evaluating dynamic queries in a real-estate information exploring system
  16. Rasmussen, E.M.: Indexing and retrieval for the Web (2002) 0.01
    0.0063776104 = product of:
      0.03826566 = sum of:
        0.03826566 = weight(_text_:searching in 4285) [ClassicSimilarity], result of:
          0.03826566 = score(doc=4285,freq=6.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.2709465 = fieldWeight in 4285, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4285)
      0.16666667 = coord(1/6)
    
    Abstract
    The introduction and growth of the World Wide Web (WWW, or Web) have resulted in a profound change in the way individuals and organizations access information. In terms of volume, nature, and accessibility, the characteristics of electronic information are significantly different from those of even five or six years ago. Control of, and access to, this flood of information rely heavily an automated techniques for indexing and retrieval. According to Gudivada, Raghavan, Grosky, and Kasanagottu (1997, p. 58), "The ability to search and retrieve information from the Web efficiently and effectively is an enabling technology for realizing its full potential." Almost 93 percent of those surveyed consider the Web an "indispensable" Internet technology, second only to e-mail (Graphie, Visualization & Usability Center, 1998). Although there are other ways of locating information an the Web (browsing or following directory structures), 85 percent of users identify Web pages by means of a search engine (Graphie, Visualization & Usability Center, 1998). A more recent study conducted by the Stanford Institute for the Quantitative Study of Society confirms the finding that searching for information is second only to e-mail as an Internet activity (Nie & Ebring, 2000, online). In fact, Nie and Ebring conclude, "... the Internet today is a giant public library with a decidedly commercial tilt. The most widespread use of the Internet today is as an information search utility for products, travel, hobbies, and general information. Virtually all users interviewed responded that they engaged in one or more of these information gathering activities."
    Techniques for automated indexing and information retrieval (IR) have been developed, tested, and refined over the past 40 years, and are well documented (see, for example, Agosti & Smeaton, 1996; BaezaYates & Ribeiro-Neto, 1999a; Frakes & Baeza-Yates, 1992; Korfhage, 1997; Salton, 1989; Witten, Moffat, & Bell, 1999). With the introduction of the Web, and the capability to index and retrieve via search engines, these techniques have been extended to a new environment. They have been adopted, altered, and in some Gases extended to include new methods. "In short, search engines are indispensable for searching the Web, they employ a variety of relatively advanced IR techniques, and there are some peculiar aspects of search engines that make searching the Web different than more conventional information retrieval" (Gordon & Pathak, 1999, p. 145). The environment for information retrieval an the World Wide Web differs from that of "conventional" information retrieval in a number of fundamental ways. The collection is very large and changes continuously, with pages being added, deleted, and altered. Wide variability between the size, structure, focus, quality, and usefulness of documents makes Web documents much more heterogeneous than a typical electronic document collection. The wide variety of document types includes images, video, audio, and scripts, as well as many different document languages. Duplication of documents and sites is common. Documents are interconnected through networks of hyperlinks. Because of the size and dynamic nature of the Web, preprocessing all documents requires considerable resources and is often not feasible, certainly not an the frequent basis required to ensure currency. Query length is usually much shorter than in other environments-only a few words-and user behavior differs from that in other environments. These differences make the Web a novel environment for information retrieval (Baeza-Yates & Ribeiro-Neto, 1999b; Bharat & Henzinger, 1998; Huang, 2000).
  17. Advances in intelligent retrieval: Proc. of a conference ... Wadham College, Oxford, 16.-17.4.1985 (1986) 0.01
    0.0063121966 = product of:
      0.03787318 = sum of:
        0.03787318 = weight(_text_:searching in 1384) [ClassicSimilarity], result of:
          0.03787318 = score(doc=1384,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.26816747 = fieldWeight in 1384, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.046875 = fieldNorm(doc=1384)
      0.16666667 = coord(1/6)
    
    Content
    Enthält die Beiträge: ADDIS, T.: Extended relational analysis: a design approach to knowledge-based systems; PARKINSON, D.: Supercomputers and non-numeric processing; McGREGOR, D.R. u. J.R. MALONE: An architectural approach to advances in information retrieval; ALLEN, M.J. u. O.S. HARRISON: Word processing and information retrieval: some practical problems; MURTAGH, F.: Clustering and nearest neighborhood searching; ENSER, P.G.B.: Experimenting with the automatic classification of books; TESKEY, N. u. Z. RAZAK: An analysis of ranking for free text retrieval systems; ZARRI, G.P.: Interactive information retrieval: an artificial intelligence approach to deal with biographical data; HANCOX, P. u. F. SMITH: A case system processor for the PRECIS indexing language; ROUAULT, J.: Linguistic methods in information retrieval systems; ARAGON-RAMIREZ, V. u. C.D. PAICE: Design of a system for the online elucidation of natural language search statements; BROOKS, H.M., P.J. DANIELS u. N.J. BELKIN: Problem descriptions and user models: developing an intelligent interface for document retrieval systems; BLACK, W.J., P. HARGREAVES u. P.B. MAYES: HEADS: a cataloguing advisory system; BELL, D.A.: An architecture for integrating data, knowledge, and information bases
  18. Wacholder, N.; Byrd, R.J.: Retrieving information from full text using linguistic knowledge (1994) 0.01
    0.0063121966 = product of:
      0.03787318 = sum of:
        0.03787318 = weight(_text_:searching in 8524) [ClassicSimilarity], result of:
          0.03787318 = score(doc=8524,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.26816747 = fieldWeight in 8524, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.046875 = fieldNorm(doc=8524)
      0.16666667 = coord(1/6)
    
    Abstract
    Examines how techniques in the field of natural language processing can be applied to the analysis of text in information retrieval. State of the art text searching programs cannot distinguish, for example, between occurrences of the sickness, AIDS and aids as tool or between library school and school nor equate such terms as online or on-line which are variants of the same form. To make these distinction, systems must incorporate knowledge about the meaning of words in context. Research in natural language processing has concentrated on the automatic 'understanding' of language; how to analyze the grammatical structure and meaning of text. Although many asoects of this research remain experimental, describes how these techniques to recognize spelling variants, names, acronyms, and abbreviations
  19. Liu, G.Z.: Semantic vector space model : implementation and evaluation (1997) 0.01
    0.0063121966 = product of:
      0.03787318 = sum of:
        0.03787318 = weight(_text_:searching in 161) [ClassicSimilarity], result of:
          0.03787318 = score(doc=161,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.26816747 = fieldWeight in 161, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.046875 = fieldNorm(doc=161)
      0.16666667 = coord(1/6)
    
    Abstract
    Presents the Semantic Vector Space Model (SVSM), a text representation and searching technique based on the combination of Vector Space Model (VSM) with heuristic syntax parsing and distributed representation of semantic case structures. Both document and queries are represented as semantic matrices. A search mechanism is designed to compute the similarity between 2 semantic matrices to predict relevancy. A prototype system was built to implement this model by modifying the SMART system and using the Xerox Part of Speech tagged as the pre-processor of the indexing. The prototype system was used in an experimental study to evaluate this technique in terms of precision, recall, and effectiveness of relevance ranking. Results show that if documents and queries were too short, the technique was less effective than VSM. But with longer documents and queires, especially when original docuemtns were used as queries, the system based on this technique was found be performance better than SMART
  20. MacDougall, S.: Rethinking indexing : the impact of the Internet (1996) 0.01
    0.0063121966 = product of:
      0.03787318 = sum of:
        0.03787318 = weight(_text_:searching in 704) [ClassicSimilarity], result of:
          0.03787318 = score(doc=704,freq=2.0), product of:
            0.14122958 = queryWeight, product of:
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.03491209 = queryNorm
            0.26816747 = fieldWeight in 704, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.0452914 = idf(docFreq=2103, maxDocs=44218)
              0.046875 = fieldNorm(doc=704)
      0.16666667 = coord(1/6)
    
    Abstract
    Considers the challenge to professional indexers posed by the Internet. Indexing and searching on the Internet appears to have a retrograde step, as well developed and efficient information retrieval techniques have been replaced by cruder techniques, involving automatic keyword indexing and frequency ranking, leading to large retrieval sets and low precision. This is made worse by the apparent acceptance of this poor perfromance by Internet users and the feeling, on the part of indexers, that they are being bypassed by the producers of these hyperlinked menus and search engines. Key issues are: how far 'human' indexing will still be required in the Internet environment; how indexing techniques will have to change to stay relevant; and the future role of indexers. The challenge facing indexers is to adapt their skills to suit the online environment and to convince publishers of the need for efficient indexes on the Internet

Years

Languages

  • e 39
  • d 17
  • f 1
  • ja 1
  • ru 1
  • More… Less…

Types

  • a 53
  • el 3
  • s 2
  • x 2
  • m 1
  • More… Less…