Search (281 results, page 1 of 15)

  • × theme_ss:"Automatisches Indexieren"
  1. Hlava, M.M.K.: Automatic indexing : comparing rule-based and statistics-based indexing systems (2005) 0.13
    0.13283804 = product of:
      0.19925706 = sum of:
        0.024771197 = weight(_text_:information in 6265) [ClassicSimilarity], result of:
          0.024771197 = score(doc=6265,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.27153665 = fieldWeight in 6265, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=6265)
        0.17448586 = sum of:
          0.07591552 = weight(_text_:systems in 6265) [ClassicSimilarity], result of:
            0.07591552 = score(doc=6265,freq=2.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.47535738 = fieldWeight in 6265, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.109375 = fieldNorm(doc=6265)
          0.09857034 = weight(_text_:22 in 6265) [ClassicSimilarity], result of:
            0.09857034 = score(doc=6265,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.5416616 = fieldWeight in 6265, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.109375 = fieldNorm(doc=6265)
      0.6666667 = coord(2/3)
    
    Source
    Information outlook. 9(2005) no.8, S.22-23
  2. Lepsky, K.; Vorhauer, J.: Lingo - ein open source System für die Automatische Indexierung deutschsprachiger Dokumente (2006) 0.08
    0.075907454 = product of:
      0.11386118 = sum of:
        0.01415497 = weight(_text_:information in 3581) [ClassicSimilarity], result of:
          0.01415497 = score(doc=3581,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.1551638 = fieldWeight in 3581, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=3581)
        0.09970621 = sum of:
          0.0433803 = weight(_text_:systems in 3581) [ClassicSimilarity], result of:
            0.0433803 = score(doc=3581,freq=2.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.2716328 = fieldWeight in 3581, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.0625 = fieldNorm(doc=3581)
          0.05632591 = weight(_text_:22 in 3581) [ClassicSimilarity], result of:
            0.05632591 = score(doc=3581,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.30952093 = fieldWeight in 3581, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=3581)
      0.6666667 = coord(2/3)
    
    Abstract
    Lingo ist ein frei verfügbares System (open source) zur automatischen Indexierung der deutschen Sprache. Bei der Entwicklung von lingo standen hohe Konfigurierbarkeit und Flexibilität des Systems für unterschiedliche Einsatzmöglichkeiten im Vordergrund. Der Beitrag zeigt den Nutzen einer linguistisch basierten automatischen Indexierung für das Information Retrieval auf. Die für eine Retrievalverbesserung zur Verfügung stehende linguistische Funktionalität von lingo wird vorgestellt und an Beispielen erläutert: Grundformerkennung, Kompositumerkennung bzw. Kompositumzerlegung, Wortrelationierung, lexikalische und algorithmische Mehrwortgruppenerkennung, OCR-Fehlerkorrektur. Der offene Systemaufbau von lingo wird beschrieben, mögliche Einsatzszenarien und Anwendungsgrenzen werden benannt.
    Date
    24. 3.2006 12:22:02
  3. Wolfekuhler, M.R.; Punch, W.F.: Finding salient features for personal Web pages categories (1997) 0.07
    0.06641902 = product of:
      0.09962853 = sum of:
        0.012385598 = weight(_text_:information in 2673) [ClassicSimilarity], result of:
          0.012385598 = score(doc=2673,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.13576832 = fieldWeight in 2673, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2673)
        0.08724293 = sum of:
          0.03795776 = weight(_text_:systems in 2673) [ClassicSimilarity], result of:
            0.03795776 = score(doc=2673,freq=2.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.23767869 = fieldWeight in 2673, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2673)
          0.04928517 = weight(_text_:22 in 2673) [ClassicSimilarity], result of:
            0.04928517 = score(doc=2673,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.2708308 = fieldWeight in 2673, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2673)
      0.6666667 = coord(2/3)
    
    Abstract
    Examines techniques that discover features in sets of pre-categorized documents, such that similar documents can be found on the WWW. Examines techniques which will classifiy training examples with high accuracy, then explains why this is not necessarily useful. Describes a method for extracting word clusters from the raw document features. Results show that the clustering technique is successful in discovering word groups in personal Web pages which can be used to find similar information on the WWW
    Date
    1. 8.1996 22:08:06
    Source
    Computer networks and ISDN systems. 29(1997) no.8, S.1147-1156
  4. Ward, M.L.: ¬The future of the human indexer (1996) 0.06
    0.056930594 = product of:
      0.08539589 = sum of:
        0.010616227 = weight(_text_:information in 7244) [ClassicSimilarity], result of:
          0.010616227 = score(doc=7244,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.116372846 = fieldWeight in 7244, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=7244)
        0.07477966 = sum of:
          0.032535225 = weight(_text_:systems in 7244) [ClassicSimilarity], result of:
            0.032535225 = score(doc=7244,freq=2.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.2037246 = fieldWeight in 7244, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.046875 = fieldNorm(doc=7244)
          0.04224443 = weight(_text_:22 in 7244) [ClassicSimilarity], result of:
            0.04224443 = score(doc=7244,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.23214069 = fieldWeight in 7244, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=7244)
      0.6666667 = coord(2/3)
    
    Abstract
    Considers the principles of indexing and the intellectual skills involved in order to determine what automatic indexing systems would be required in order to supplant or complement the human indexer. Good indexing requires: considerable prior knowledge of the literature; judgement as to what to index and what depth to index; reading skills; abstracting skills; and classification skills, Illustrates these features with a detailed description of abstracting and indexing processes involved in generating entries for the mechanical engineering database POWERLINK. Briefly assesses the possibility of replacing human indexers with specialist indexing software, with particular reference to the Object Analyzer from the InTEXT automatic indexing system and using the criteria described for human indexers. At present, it is unlikely that the automatic indexer will replace the human indexer, but when more primary texts are available in electronic form, it may be a useful productivity tool for dealing with large quantities of low grade texts (should they be wanted in the database)
    Date
    9. 2.1997 18:44:22
    Source
    Journal of librarianship and information science. 28(1996) no.4, S.217-225
  5. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.06
    0.0564239 = product of:
      0.08463585 = sum of:
        0.02830994 = weight(_text_:information in 402) [ClassicSimilarity], result of:
          0.02830994 = score(doc=402,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3103276 = fieldWeight in 402, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=402)
        0.05632591 = product of:
          0.11265182 = sum of:
            0.11265182 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
              0.11265182 = score(doc=402,freq=2.0), product of:
                0.1819777 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051966466 = queryNorm
                0.61904186 = fieldWeight in 402, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=402)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
  6. Greiner-Petter, A.; Schubotz, M.; Cohl, H.S.; Gipp, B.: Semantic preserving bijective mappings for expressions involving special functions between computer algebra systems and document preparation systems (2019) 0.05
    0.05436822 = product of:
      0.08155233 = sum of:
        0.010009075 = weight(_text_:information in 5499) [ClassicSimilarity], result of:
          0.010009075 = score(doc=5499,freq=4.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.10971737 = fieldWeight in 5499, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=5499)
        0.071543254 = sum of:
          0.0433803 = weight(_text_:systems in 5499) [ClassicSimilarity], result of:
            0.0433803 = score(doc=5499,freq=8.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.2716328 = fieldWeight in 5499, product of:
                2.828427 = tf(freq=8.0), with freq of:
                  8.0 = termFreq=8.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.03125 = fieldNorm(doc=5499)
          0.028162954 = weight(_text_:22 in 5499) [ClassicSimilarity], result of:
            0.028162954 = score(doc=5499,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.15476047 = fieldWeight in 5499, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=5499)
      0.6666667 = coord(2/3)
    
    Abstract
    Purpose Modern mathematicians and scientists of math-related disciplines often use Document Preparation Systems (DPS) to write and Computer Algebra Systems (CAS) to calculate mathematical expressions. Usually, they translate the expressions manually between DPS and CAS. This process is time-consuming and error-prone. The purpose of this paper is to automate this translation. This paper uses Maple and Mathematica as the CAS, and LaTeX as the DPS. Design/methodology/approach Bruce Miller at the National Institute of Standards and Technology (NIST) developed a collection of special LaTeX macros that create links from mathematical symbols to their definitions in the NIST Digital Library of Mathematical Functions (DLMF). The authors are using these macros to perform rule-based translations between the formulae in the DLMF and CAS. Moreover, the authors develop software to ease the creation of new rules and to discover inconsistencies. Findings The authors created 396 mappings and translated 58.8 percent of DLMF formulae (2,405 expressions) successfully between Maple and DLMF. For a significant percentage, the special function definitions in Maple and the DLMF were different. An atomic symbol in one system maps to a composite expression in the other system. The translator was also successfully used for automatic verification of mathematical online compendia and CAS. The evaluation techniques discovered two errors in the DLMF and one defect in Maple. Originality/value This paper introduces the first translation tool for special functions between LaTeX and CAS. The approach improves error-prone manual translations and can be used to verify mathematical online compendia and CAS.
    Date
    20. 1.2015 18:30:22
    Footnote
    Beitrag in einem Special Issue: Information Science in the German-speaking Countries.
    Source
    Aslib journal of information management. 71(2019) no.3, S.415-439
  7. Milstead, J.L.: Thesauri in a full-text world (1998) 0.05
    0.053340062 = product of:
      0.080010094 = sum of:
        0.017693711 = weight(_text_:information in 2337) [ClassicSimilarity], result of:
          0.017693711 = score(doc=2337,freq=8.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.19395474 = fieldWeight in 2337, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2337)
        0.06231638 = sum of:
          0.027112689 = weight(_text_:systems in 2337) [ClassicSimilarity], result of:
            0.027112689 = score(doc=2337,freq=2.0), product of:
              0.159702 = queryWeight, product of:
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.051966466 = queryNorm
              0.1697705 = fieldWeight in 2337, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.0731742 = idf(docFreq=5561, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2337)
          0.03520369 = weight(_text_:22 in 2337) [ClassicSimilarity], result of:
            0.03520369 = score(doc=2337,freq=2.0), product of:
              0.1819777 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051966466 = queryNorm
              0.19345059 = fieldWeight in 2337, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2337)
      0.6666667 = coord(2/3)
    
    Abstract
    Despite early claims to the contemporary, thesauri continue to find use as access tools for information in the full-text environment. Their mode of use is changing, but this change actually represents an expansion rather than a contrdiction of their utility. Thesauri and similar vocabulary tools can complement full-text access by aiding users in focusing their searches, by supplementing the linguistic analysis of the text search engine, and even by serving as one of the tools used by the linguistic engine for its analysis. While human indexing contunues to be used for many databases, the trend is to increase the use of machine aids for this purpose. All machine-aided indexing (MAI) systems rely on thesauri as the basis for term selection. In the 21st century, the balance of effort between human and machine will change at both input and output, but thesauri will continue to play an important role for the foreseeable future
    Date
    22. 9.1997 19:16:05
    Imprint
    Urbana-Champaign, IL : Illinois University at Urbana-Champaign, Graduate School of Library and Information Science
    Source
    Visualizing subject access for 21st century information resources: Papers presented at the 1997 Clinic on Library Applications of Data Processing, 2-4 Mar 1997, Graduate School of Library and Information Science, University of Illinois at Urbana-Champaign. Ed.: P.A. Cochrane et al
  8. Griffiths, A.; Luckhurst, H.C.; Willett, P.: Using interdocument similarity information in document retrieval systems (1986) 0.05
    0.048659682 = product of:
      0.07298952 = sum of:
        0.035031762 = weight(_text_:information in 2415) [ClassicSimilarity], result of:
          0.035031762 = score(doc=2415,freq=4.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3840108 = fieldWeight in 2415, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=2415)
        0.03795776 = product of:
          0.07591552 = sum of:
            0.07591552 = weight(_text_:systems in 2415) [ClassicSimilarity], result of:
              0.07591552 = score(doc=2415,freq=2.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.47535738 = fieldWeight in 2415, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.109375 = fieldNorm(doc=2415)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Journal of the American Society for Information Science. 37(1986) no.1, S.3-11
  9. Thiel, T.J.: Automated indexing of information stored on optical disk electronic document image management systems (1994) 0.05
    0.048659682 = product of:
      0.07298952 = sum of:
        0.035031762 = weight(_text_:information in 1260) [ClassicSimilarity], result of:
          0.035031762 = score(doc=1260,freq=4.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3840108 = fieldWeight in 1260, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.109375 = fieldNorm(doc=1260)
        0.03795776 = product of:
          0.07591552 = sum of:
            0.07591552 = weight(_text_:systems in 1260) [ClassicSimilarity], result of:
              0.07591552 = score(doc=1260,freq=2.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.47535738 = fieldWeight in 1260, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.109375 = fieldNorm(doc=1260)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Encyclopedia of library and information science. Vol.54, [=Suppl.17]
  10. Smart, G.: Using language analysis to manage information (1993) 0.04
    0.041550655 = product of:
      0.06232598 = sum of:
        0.031651475 = weight(_text_:information in 4423) [ClassicSimilarity], result of:
          0.031651475 = score(doc=4423,freq=10.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3469568 = fieldWeight in 4423, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=4423)
        0.030674506 = product of:
          0.061349012 = sum of:
            0.061349012 = weight(_text_:systems in 4423) [ClassicSimilarity], result of:
              0.061349012 = score(doc=4423,freq=4.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.38414678 = fieldWeight in 4423, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4423)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The ESPRIT project SIMPR developed software to analyse documents and generate indexes for them. Of immediate application as a document indexing and classification system, this also offers a technology for information modelling that has broader implications, supporting many new uses for information management softeware. The project was based on the assumption that information can only be managed successfully by computer systems that can view the information contained in a document through the language in which the document is written, and that systems need to be sufficiently flexible to respond to the changing requirements of document use
  11. Biebricher, N.; Fuhr, N.; Lustig, G.; Schwantner, M.; Knorz, G.: ¬The automatic indexing system AIR/PHYS : from research to application (1988) 0.04
    0.04015092 = product of:
      0.06022638 = sum of:
        0.025022687 = weight(_text_:information in 1952) [ClassicSimilarity], result of:
          0.025022687 = score(doc=1952,freq=4.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.27429342 = fieldWeight in 1952, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=1952)
        0.03520369 = product of:
          0.07040738 = sum of:
            0.07040738 = weight(_text_:22 in 1952) [ClassicSimilarity], result of:
              0.07040738 = score(doc=1952,freq=2.0), product of:
                0.1819777 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051966466 = queryNorm
                0.38690117 = fieldWeight in 1952, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1952)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Date
    16. 8.1998 12:51:22
    Footnote
    Wiederabgedruckt in: Readings in information retrieval. Ed.: K. Sparck Jones u. P. Willett. San Francisco: Morgan Kaufmann 1997. S.513-517.
    Source
    Proceedings of the 11th annual conference on research and development in information retrieval. Ed.: Y. Chiaramella
  12. Paijmans, H.: Comparing the document representation of two IR-systems : CLARIT and TOPIC (1993) 0.04
    0.03839106 = product of:
      0.05758659 = sum of:
        0.02001815 = weight(_text_:information in 6503) [ClassicSimilarity], result of:
          0.02001815 = score(doc=6503,freq=4.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.21943474 = fieldWeight in 6503, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=6503)
        0.037568443 = product of:
          0.075136885 = sum of:
            0.075136885 = weight(_text_:systems in 6503) [ClassicSimilarity], result of:
              0.075136885 = score(doc=6503,freq=6.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.4704818 = fieldWeight in 6503, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6503)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Discusses the TOPIC and CLARIT information retrieval systems in terms of assigned versus derived and precoordinate versus postcoordinate indexing. Compares the document representation of the two systems. Reports on a test done on a small sample of Wall Street Journal articles. The positive results found for CLARIT in earlier test on medical documents were not observed in this general database
    Source
    Journal of the American Society for Information Science. 44(1993) no.7, S.383-392
  13. Advances in intelligent retrieval: Proc. of a conference ... Wadham College, Oxford, 16.-17.4.1985 (1986) 0.04
    0.03751589 = product of:
      0.056273833 = sum of:
        0.023738606 = weight(_text_:information in 1384) [ClassicSimilarity], result of:
          0.023738606 = score(doc=1384,freq=10.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.2602176 = fieldWeight in 1384, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=1384)
        0.032535225 = product of:
          0.06507045 = sum of:
            0.06507045 = weight(_text_:systems in 1384) [ClassicSimilarity], result of:
              0.06507045 = score(doc=1384,freq=8.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.4074492 = fieldWeight in 1384, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1384)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Content
    Enthält die Beiträge: ADDIS, T.: Extended relational analysis: a design approach to knowledge-based systems; PARKINSON, D.: Supercomputers and non-numeric processing; McGREGOR, D.R. u. J.R. MALONE: An architectural approach to advances in information retrieval; ALLEN, M.J. u. O.S. HARRISON: Word processing and information retrieval: some practical problems; MURTAGH, F.: Clustering and nearest neighborhood searching; ENSER, P.G.B.: Experimenting with the automatic classification of books; TESKEY, N. u. Z. RAZAK: An analysis of ranking for free text retrieval systems; ZARRI, G.P.: Interactive information retrieval: an artificial intelligence approach to deal with biographical data; HANCOX, P. u. F. SMITH: A case system processor for the PRECIS indexing language; ROUAULT, J.: Linguistic methods in information retrieval systems; ARAGON-RAMIREZ, V. u. C.D. PAICE: Design of a system for the online elucidation of natural language search statements; BROOKS, H.M., P.J. DANIELS u. N.J. BELKIN: Problem descriptions and user models: developing an intelligent interface for document retrieval systems; BLACK, W.J., P. HARGREAVES u. P.B. MAYES: HEADS: a cataloguing advisory system; BELL, D.A.: An architecture for integrating data, knowledge, and information bases
  14. Kutschekmanesch, S.; Lutes, B.; Moelle, K.; Thiel, U.; Tzeras, K.: Automated multilingual indexing : a synthesis of rule-based and thesaurus-based methods (1998) 0.04
    0.035264935 = product of:
      0.0528974 = sum of:
        0.017693711 = weight(_text_:information in 4157) [ClassicSimilarity], result of:
          0.017693711 = score(doc=4157,freq=2.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.19395474 = fieldWeight in 4157, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.078125 = fieldNorm(doc=4157)
        0.03520369 = product of:
          0.07040738 = sum of:
            0.07040738 = weight(_text_:22 in 4157) [ClassicSimilarity], result of:
              0.07040738 = score(doc=4157,freq=2.0), product of:
                0.1819777 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051966466 = queryNorm
                0.38690117 = fieldWeight in 4157, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=4157)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Information und Märkte: 50. Deutscher Dokumentartag 1998, Kongreß der Deutschen Gesellschaft für Dokumentation e.V. (DGD), Rheinische Friedrich-Wilhelms-Universität Bonn, 22.-24. September 1998. Hrsg. von Marlies Ockenfeld u. Gerhard J. Mantwill
  15. Riloff, E.: ¬An empirical study of automated dictionary construction for information extraction in three domains (1996) 0.04
    0.035120055 = product of:
      0.052680083 = sum of:
        0.024517128 = weight(_text_:information in 6752) [ClassicSimilarity], result of:
          0.024517128 = score(doc=6752,freq=6.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.2687516 = fieldWeight in 6752, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=6752)
        0.028162954 = product of:
          0.05632591 = sum of:
            0.05632591 = weight(_text_:22 in 6752) [ClassicSimilarity], result of:
              0.05632591 = score(doc=6752,freq=2.0), product of:
                0.1819777 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051966466 = queryNorm
                0.30952093 = fieldWeight in 6752, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6752)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    AutoSlog is a system that addresses the knowledge engineering bottleneck for information extraction. AutoSlog automatically creates domain specific dictionaries for information extraction, given an appropriate training corpus. Describes experiments with AutoSlog in terrorism, joint ventures and microelectronics domains. Compares the performance of AutoSlog across the 3 domains, discusses the lessons learned and presents results from 2 experiments which demonstrate that novice users can generate effective dictionaries using AutoSlog
    Date
    6. 3.1997 16:22:15
  16. Kim, P.K.: ¬An automatic indexing of compound words based on mutual information for Korean text retrieval (1995) 0.03
    0.033333395 = product of:
      0.05000009 = sum of:
        0.02830994 = weight(_text_:information in 620) [ClassicSimilarity], result of:
          0.02830994 = score(doc=620,freq=8.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3103276 = fieldWeight in 620, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=620)
        0.02169015 = product of:
          0.0433803 = sum of:
            0.0433803 = weight(_text_:systems in 620) [ClassicSimilarity], result of:
              0.0433803 = score(doc=620,freq=2.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.2716328 = fieldWeight in 620, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0625 = fieldNorm(doc=620)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Presents an automatic indexing technique for compound words suitable for an agglutinative language, specifically Korean. Discusses some construction conditions for compound words and the rules for decomposing compound words to enhance the exhaustivity of indexing, demonstrating that this system, mutual information, enhances both the exhaustivity of indexing and the specifity of terms. Suggests that the construction conditions and rules for decomposition presented may be used in multilingual information retrieval systems to translate the indexing terms of the specific language into those of the language required
    Source
    Library and information science. 1995, no.34, S.29-38
  17. Losee, R.M.: ¬A Gray code based ordering for documents on shelves : classification for browsing and retrieval (1992) 0.03
    0.03219512 = product of:
      0.04829268 = sum of:
        0.021452487 = weight(_text_:information in 2335) [ClassicSimilarity], result of:
          0.021452487 = score(doc=2335,freq=6.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.23515764 = fieldWeight in 2335, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2335)
        0.026840193 = product of:
          0.053680386 = sum of:
            0.053680386 = weight(_text_:systems in 2335) [ClassicSimilarity], result of:
              0.053680386 = score(doc=2335,freq=4.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.33612844 = fieldWeight in 2335, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2335)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    A document classifier places documents together in a linear arrangement for browsing or high-speed access by human or computerised information retrieval systems. Requirements for document classification and browsing systems are developed from similarity measures, distance measures, and the notion of subject aboutness. A requirement that documents be arranged in decreasing order of similarity as the distance from a given document increases can often not be met. Based on these requirements, information-theoretic considerations, and the Gray code, a classification system is proposed that can classifiy documents without human intervention. A measure of classifier performance is developed, and used to evaluate experimental results comparing the distance between subject headings assigned to documents given classifications from the proposed system and the Library of Congress Classification (LCC) system
    Source
    Journal of the American Society for Information Science. 43(1992) no.4, S.312-322
  18. Hersh, W.R.; Hickam, D.H.: ¬A comparison of two methods for indexing and retrieval from a full-text medical database (1992) 0.03
    0.03219512 = product of:
      0.04829268 = sum of:
        0.021452487 = weight(_text_:information in 4526) [ClassicSimilarity], result of:
          0.021452487 = score(doc=4526,freq=6.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.23515764 = fieldWeight in 4526, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4526)
        0.026840193 = product of:
          0.053680386 = sum of:
            0.053680386 = weight(_text_:systems in 4526) [ClassicSimilarity], result of:
              0.053680386 = score(doc=4526,freq=4.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.33612844 = fieldWeight in 4526, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4526)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Reports results of a study of 2 information retrieval systems on a 2.000 document full text medical database. The first system, SAPHIRE, features concept based automatic indexing and statistical retrieval techniques, while the second system, SWORD, features traditional word based Boolean techniques, 16 medical students at Oregon Health Sciences Univ. each performed 10 searches and their results, recorded in terms of recall and precision, showed nearly equal performance for both systems. SAPHIRE was also compared with a version of SWORD modified to use automatic indexing and ranked retrieval. Using batch input of queries, the latter method performed slightly better
    Imprint
    Medford, NJ : Learned Information Inc.
    Source
    Proceedings of the 55th Annual Meeting of the American Society for Information Science, Pittsburgh, 26.-29.10.92. Ed.: D. Shaw
  19. Krutulis, J.D.; Jacob, E.K.: ¬A theoretical model for the study of emergent structure in adaptive information networks (1995) 0.03
    0.031115947 = product of:
      0.04667392 = sum of:
        0.02769504 = weight(_text_:information in 3353) [ClassicSimilarity], result of:
          0.02769504 = score(doc=3353,freq=10.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.3035872 = fieldWeight in 3353, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3353)
        0.01897888 = product of:
          0.03795776 = sum of:
            0.03795776 = weight(_text_:systems in 3353) [ClassicSimilarity], result of:
              0.03795776 = score(doc=3353,freq=2.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.23767869 = fieldWeight in 3353, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3353)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Attempts to automate classification have focused on mimicking the intellectual processes whereby human classifiers assign entities to mutually exclusive groups that exhibit or more shared characteristics. A more viable approach might be to construct an adaptive retrieval system that produces groupings of related entities by generating dynamic categories based on document content and on the system's emergent structure as it adapts to modifications in the database and to observed patterns of access. Presents a theoretical model for adaptive information networks using relevance feedback and genetic algorithms to generate emergent structure
    Imprint
    Alberta : Alberta University, School of Library and Information Studies
    Source
    Connectedness: information, systems, people, organizations. Proceedings of CAIS/ACSI 95, the proceedings of the 23rd Annual Conference of the Canadian Association for Information Science. Ed. by Hope A. Olson and Denis B. Ward
  20. Mars, N.J.I.: ¬The management of scientific information, or, how to cope with the flood (1996) 0.03
    0.030804854 = product of:
      0.04620728 = sum of:
        0.024517128 = weight(_text_:information in 7414) [ClassicSimilarity], result of:
          0.024517128 = score(doc=7414,freq=6.0), product of:
            0.09122598 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.051966466 = queryNorm
            0.2687516 = fieldWeight in 7414, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=7414)
        0.02169015 = product of:
          0.0433803 = sum of:
            0.0433803 = weight(_text_:systems in 7414) [ClassicSimilarity], result of:
              0.0433803 = score(doc=7414,freq=2.0), product of:
                0.159702 = queryWeight, product of:
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.051966466 = queryNorm
                0.2716328 = fieldWeight in 7414, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.0731742 = idf(docFreq=5561, maxDocs=44218)
                  0.0625 = fieldNorm(doc=7414)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Research in the Knowledge-Based Systems Group of the University of Twente in the Netherlands is aimed at reducing information overload. One approach is to support indexing by the traditional method of assigning content descriptions to find documents. A second way is to use a computer program to determine what the document says without descriptors. Discusses automated indexing and direct access to information

Languages

Types

  • a 244
  • x 14
  • el 13
  • m 13
  • s 8
  • d 1
  • More… Less…