Search (109 results, page 1 of 6)

  • × theme_ss:"Automatisches Indexieren"
  1. Biebricher, N.; Fuhr, N.; Lustig, G.; Schwantner, M.; Knorz, G.: ¬The automatic indexing system AIR/PHYS : from research to application (1988) 0.05
    0.05112489 = product of:
      0.12781222 = sum of:
        0.057155732 = weight(_text_:system in 1952) [ClassicSimilarity], result of:
          0.057155732 = score(doc=1952,freq=2.0), product of:
            0.1642502 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.052150324 = queryNorm
            0.3479797 = fieldWeight in 1952, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.078125 = fieldNorm(doc=1952)
        0.07065649 = weight(_text_:22 in 1952) [ClassicSimilarity], result of:
          0.07065649 = score(doc=1952,freq=2.0), product of:
            0.18262155 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052150324 = queryNorm
            0.38690117 = fieldWeight in 1952, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.078125 = fieldNorm(doc=1952)
      0.4 = coord(2/5)
    
    Date
    16. 8.1998 12:51:22
  2. Lepsky, K.; Vorhauer, J.: Lingo - ein open source System für die Automatische Indexierung deutschsprachiger Dokumente (2006) 0.05
    0.048475813 = product of:
      0.12118953 = sum of:
        0.064664334 = weight(_text_:system in 3581) [ClassicSimilarity], result of:
          0.064664334 = score(doc=3581,freq=4.0), product of:
            0.1642502 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.052150324 = queryNorm
            0.3936941 = fieldWeight in 3581, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0625 = fieldNorm(doc=3581)
        0.056525193 = weight(_text_:22 in 3581) [ClassicSimilarity], result of:
          0.056525193 = score(doc=3581,freq=2.0), product of:
            0.18262155 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052150324 = queryNorm
            0.30952093 = fieldWeight in 3581, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0625 = fieldNorm(doc=3581)
      0.4 = coord(2/5)
    
    Abstract
    Lingo ist ein frei verfügbares System (open source) zur automatischen Indexierung der deutschen Sprache. Bei der Entwicklung von lingo standen hohe Konfigurierbarkeit und Flexibilität des Systems für unterschiedliche Einsatzmöglichkeiten im Vordergrund. Der Beitrag zeigt den Nutzen einer linguistisch basierten automatischen Indexierung für das Information Retrieval auf. Die für eine Retrievalverbesserung zur Verfügung stehende linguistische Funktionalität von lingo wird vorgestellt und an Beispielen erläutert: Grundformerkennung, Kompositumerkennung bzw. Kompositumzerlegung, Wortrelationierung, lexikalische und algorithmische Mehrwortgruppenerkennung, OCR-Fehlerkorrektur. Der offene Systemaufbau von lingo wird beschrieben, mögliche Einsatzszenarien und Anwendungsgrenzen werden benannt.
    Date
    24. 3.2006 12:22:02
  3. Bordoni, L.; Pazienza, M.T.: Documents automatic indexing in an environmental domain (1997) 0.04
    0.042416334 = product of:
      0.106040835 = sum of:
        0.05658129 = weight(_text_:system in 530) [ClassicSimilarity], result of:
          0.05658129 = score(doc=530,freq=4.0), product of:
            0.1642502 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.052150324 = queryNorm
            0.34448233 = fieldWeight in 530, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0546875 = fieldNorm(doc=530)
        0.049459543 = weight(_text_:22 in 530) [ClassicSimilarity], result of:
          0.049459543 = score(doc=530,freq=2.0), product of:
            0.18262155 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052150324 = queryNorm
            0.2708308 = fieldWeight in 530, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0546875 = fieldNorm(doc=530)
      0.4 = coord(2/5)
    
    Abstract
    Describes an application of Natural Language Processing (NLP) techniques, in HIRMA (Hypertextual Information Retrieval Managed by ARIOSTO), to the problem of document indexing by referring to a system which incorporates natural language processing techniques to determine the subject of the text of documents and to associate them with relevant semantic indexes. Describes briefly the overall system, details of its implementation on a corpus of scientific abstracts related to environmental topics and experimental evidence of the system's behaviour. Analyzes in detail an experiment designed to evaluate the system's retrieval ability in terms of recall and precision
    Source
    International forum on information and documentation. 22(1997) no.1, S.17-28
  4. Tsujii, J.-I.: Automatic acquisition of semantic collocation from corpora (1995) 0.04
    0.040899914 = product of:
      0.10224978 = sum of:
        0.045724586 = weight(_text_:system in 4709) [ClassicSimilarity], result of:
          0.045724586 = score(doc=4709,freq=2.0), product of:
            0.1642502 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.052150324 = queryNorm
            0.27838376 = fieldWeight in 4709, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0625 = fieldNorm(doc=4709)
        0.056525193 = weight(_text_:22 in 4709) [ClassicSimilarity], result of:
          0.056525193 = score(doc=4709,freq=2.0), product of:
            0.18262155 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052150324 = queryNorm
            0.30952093 = fieldWeight in 4709, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0625 = fieldNorm(doc=4709)
      0.4 = coord(2/5)
    
    Abstract
    Proposes automatic linguistic knowledge acquisition from sublanguage corpora. The system combines existing linguistic knowledge and human intervention with corpus based techniques. The algorithm involves a gradual approximation which works to converge linguistic knowledge gradually towards desirable results. The 1st experiment revealed the characteristic of this algorithm and the others proved the effectiveness of this algorithm for a real corpus
    Date
    31. 7.1996 9:22:19
  5. Riloff, E.: ¬An empirical study of automated dictionary construction for information extraction in three domains (1996) 0.04
    0.040899914 = product of:
      0.10224978 = sum of:
        0.045724586 = weight(_text_:system in 6752) [ClassicSimilarity], result of:
          0.045724586 = score(doc=6752,freq=2.0), product of:
            0.1642502 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.052150324 = queryNorm
            0.27838376 = fieldWeight in 6752, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0625 = fieldNorm(doc=6752)
        0.056525193 = weight(_text_:22 in 6752) [ClassicSimilarity], result of:
          0.056525193 = score(doc=6752,freq=2.0), product of:
            0.18262155 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052150324 = queryNorm
            0.30952093 = fieldWeight in 6752, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0625 = fieldNorm(doc=6752)
      0.4 = coord(2/5)
    
    Abstract
    AutoSlog is a system that addresses the knowledge engineering bottleneck for information extraction. AutoSlog automatically creates domain specific dictionaries for information extraction, given an appropriate training corpus. Describes experiments with AutoSlog in terrorism, joint ventures and microelectronics domains. Compares the performance of AutoSlog across the 3 domains, discusses the lessons learned and presents results from 2 experiments which demonstrate that novice users can generate effective dictionaries using AutoSlog
    Date
    6. 3.1997 16:22:15
  6. Ward, M.L.: ¬The future of the human indexer (1996) 0.03
    0.030674934 = product of:
      0.076687336 = sum of:
        0.034293443 = weight(_text_:system in 7244) [ClassicSimilarity], result of:
          0.034293443 = score(doc=7244,freq=2.0), product of:
            0.1642502 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.052150324 = queryNorm
            0.20878783 = fieldWeight in 7244, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.046875 = fieldNorm(doc=7244)
        0.042393893 = weight(_text_:22 in 7244) [ClassicSimilarity], result of:
          0.042393893 = score(doc=7244,freq=2.0), product of:
            0.18262155 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052150324 = queryNorm
            0.23214069 = fieldWeight in 7244, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.046875 = fieldNorm(doc=7244)
      0.4 = coord(2/5)
    
    Abstract
    Considers the principles of indexing and the intellectual skills involved in order to determine what automatic indexing systems would be required in order to supplant or complement the human indexer. Good indexing requires: considerable prior knowledge of the literature; judgement as to what to index and what depth to index; reading skills; abstracting skills; and classification skills, Illustrates these features with a detailed description of abstracting and indexing processes involved in generating entries for the mechanical engineering database POWERLINK. Briefly assesses the possibility of replacing human indexers with specialist indexing software, with particular reference to the Object Analyzer from the InTEXT automatic indexing system and using the criteria described for human indexers. At present, it is unlikely that the automatic indexer will replace the human indexer, but when more primary texts are available in electronic form, it may be a useful productivity tool for dealing with large quantities of low grade texts (should they be wanted in the database)
    Date
    9. 2.1997 18:44:22
  7. Busch, D.: Domänenspezifische hybride automatische Indexierung von bibliographischen Metadaten (2019) 0.03
    0.030674934 = product of:
      0.076687336 = sum of:
        0.034293443 = weight(_text_:system in 5628) [ClassicSimilarity], result of:
          0.034293443 = score(doc=5628,freq=2.0), product of:
            0.1642502 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.052150324 = queryNorm
            0.20878783 = fieldWeight in 5628, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.046875 = fieldNorm(doc=5628)
        0.042393893 = weight(_text_:22 in 5628) [ClassicSimilarity], result of:
          0.042393893 = score(doc=5628,freq=2.0), product of:
            0.18262155 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052150324 = queryNorm
            0.23214069 = fieldWeight in 5628, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.046875 = fieldNorm(doc=5628)
      0.4 = coord(2/5)
    
    Abstract
    Im Fraunhofer-Informationszentrum Raum und Bau (IRB) wird Fachliteratur im Bereich Planen und Bauen bibliographisch erschlossen. Die daraus resultierenden Dokumente (Metadaten-Einträge) werden u.a. bei der Produktion der bibliographischen Datenbanken des IRB verwendet. In Abb. 1 ist ein Dokument dargestellt, das einen Zeitschriftenartikel beschreibt. Die Dokumente werden mit Deskriptoren von einer Nomenklatur (Schlagwortliste IRB) indexiert. Ein Deskriptor ist "eine Benennung., die für sich allein verwendbar, eindeutig zur Inhaltskennzeichnung geeignet und im betreffenden Dokumentationssystem zugelassen ist". Momentan wird die Indexierung intellektuell von menschlichen Experten durchgeführt. Die intellektuelle Indexierung ist zeitaufwendig und teuer. Eine Lösung des Problems besteht in der automatischen Indexierung, bei der die Zuordnung von Deskriptoren durch ein Computerprogramm erfolgt. Solche Computerprogramme werden im Folgenden auch als Klassifikatoren bezeichnet. In diesem Beitrag geht es um ein System zur automatischen Indexierung von deutschsprachigen Dokumenten im Bereich Bauwesen mit Deskriptoren aus der Schlagwortliste IRB.
    Source
    B.I.T.online. 22(2019) H.6, S.465-469
  8. Plaunt, C.; Norgard, B.A.: ¬An association-based method for automatic indexing with a controlled vocabulary (1998) 0.03
    0.025562445 = product of:
      0.06390611 = sum of:
        0.028577866 = weight(_text_:system in 1794) [ClassicSimilarity], result of:
          0.028577866 = score(doc=1794,freq=2.0), product of:
            0.1642502 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.052150324 = queryNorm
            0.17398985 = fieldWeight in 1794, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1794)
        0.035328247 = weight(_text_:22 in 1794) [ClassicSimilarity], result of:
          0.035328247 = score(doc=1794,freq=2.0), product of:
            0.18262155 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052150324 = queryNorm
            0.19345059 = fieldWeight in 1794, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1794)
      0.4 = coord(2/5)
    
    Abstract
    In this article, we describe and test a two-stage algorithm based on a lexical collocation technique which maps from the lexical clues contained in a document representation into a controlled vocabulary list of subject headings. Using a collection of 4.626 INSPEC documents, we create a 'dictionary' of associations between the lexical items contained in the titles, authors, and abstracts, and controlled vocabulary subject headings assigned to those records by human indexers using a likelihood ratio statistic as the measure of association. In the deployment stage, we use the dictiony to predict which of the controlled vocabulary subject headings best describe new documents when they are presented to the system. Our evaluation of this algorithm, in which we compare the automatically assigned subject headings to the subject headings assigned to the test documents by human catalogers, shows that we can obtain results comparable to, and consistent with, human cataloging. In effect we have cast this as a classic partial match information retrieval problem. We consider the problem to be one of 'retrieving' (or assigning) the most probably 'relevant' (or correct) controlled vocabulary subject headings to a document based on the clues contained in that document
    Date
    11. 9.2000 19:53:22
  9. Martins, A.L.; Souza, R.R.; Ribeiro de Mello, H.: ¬The use of noun phrases in information retrieval : proposing a mechanism for automatic classification (2014) 0.02
    0.024237907 = product of:
      0.060594764 = sum of:
        0.032332167 = weight(_text_:system in 1441) [ClassicSimilarity], result of:
          0.032332167 = score(doc=1441,freq=4.0), product of:
            0.1642502 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.052150324 = queryNorm
            0.19684705 = fieldWeight in 1441, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03125 = fieldNorm(doc=1441)
        0.028262597 = weight(_text_:22 in 1441) [ClassicSimilarity], result of:
          0.028262597 = score(doc=1441,freq=2.0), product of:
            0.18262155 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052150324 = queryNorm
            0.15476047 = fieldWeight in 1441, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.03125 = fieldNorm(doc=1441)
      0.4 = coord(2/5)
    
    Abstract
    This paper presents a research on syntactic structures known as noun phrases (NP) being applied to increase the effectiveness and efficiency of the mechanisms for the document's classification. Our hypothesis is the fact that the NP can be used instead of single words as a semantic aggregator to reduce the number of words that will be used for the classification system without losing its semantic coverage, increasing its efficiency. The experiment divided the documents classification process in three phases: a) NP preprocessing b) system training; and c) classification experiments. In the first step, a corpus of digitalized texts was submitted to a natural language processing platform1 in which the part-of-speech tagging was done, and them PERL scripts pertaining to the PALAVRAS package were used to extract the Noun Phrases. The preprocessing also involved the tasks of a) removing NP low meaning pre-modifiers, as quantifiers; b) identification of synonyms and corresponding substitution for common hyperonyms; and c) stemming of the relevant words contained in the NP, for similitude checking with other NPs. The first tests with the resulting documents have demonstrated its effectiveness. We have compared the structural similarity of the documents before and after the whole pre-processing steps of phase one. The texts maintained the consistency with the original and have kept the readability. The second phase involves submitting the modified documents to a SVM algorithm to identify clusters and classify the documents. The classification rules are to be established using a machine learning approach. Finally, tests will be conducted to check the effectiveness of the whole process.
    Source
    Knowledge organization in the 21st century: between historical patterns and future prospects. Proceedings of the Thirteenth International ISKO Conference 19-22 May 2014, Kraków, Poland. Ed.: Wieslaw Babik
  10. Greiner-Petter, A.; Schubotz, M.; Cohl, H.S.; Gipp, B.: Semantic preserving bijective mappings for expressions involving special functions between computer algebra systems and document preparation systems (2019) 0.02
    0.024237907 = product of:
      0.060594764 = sum of:
        0.032332167 = weight(_text_:system in 5499) [ClassicSimilarity], result of:
          0.032332167 = score(doc=5499,freq=4.0), product of:
            0.1642502 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.052150324 = queryNorm
            0.19684705 = fieldWeight in 5499, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.03125 = fieldNorm(doc=5499)
        0.028262597 = weight(_text_:22 in 5499) [ClassicSimilarity], result of:
          0.028262597 = score(doc=5499,freq=2.0), product of:
            0.18262155 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052150324 = queryNorm
            0.15476047 = fieldWeight in 5499, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.03125 = fieldNorm(doc=5499)
      0.4 = coord(2/5)
    
    Abstract
    Purpose Modern mathematicians and scientists of math-related disciplines often use Document Preparation Systems (DPS) to write and Computer Algebra Systems (CAS) to calculate mathematical expressions. Usually, they translate the expressions manually between DPS and CAS. This process is time-consuming and error-prone. The purpose of this paper is to automate this translation. This paper uses Maple and Mathematica as the CAS, and LaTeX as the DPS. Design/methodology/approach Bruce Miller at the National Institute of Standards and Technology (NIST) developed a collection of special LaTeX macros that create links from mathematical symbols to their definitions in the NIST Digital Library of Mathematical Functions (DLMF). The authors are using these macros to perform rule-based translations between the formulae in the DLMF and CAS. Moreover, the authors develop software to ease the creation of new rules and to discover inconsistencies. Findings The authors created 396 mappings and translated 58.8 percent of DLMF formulae (2,405 expressions) successfully between Maple and DLMF. For a significant percentage, the special function definitions in Maple and the DLMF were different. An atomic symbol in one system maps to a composite expression in the other system. The translator was also successfully used for automatic verification of mathematical online compendia and CAS. The evaluation techniques discovered two errors in the DLMF and one defect in Maple. Originality/value This paper introduces the first translation tool for special functions between LaTeX and CAS. The approach improves error-prone manual translations and can be used to verify mathematical online compendia and CAS.
    Date
    20. 1.2015 18:30:22
  11. Lassalle, E.: Text retrieval : from a monolingual system to a multilingual system (1993) 0.02
    0.022632515 = product of:
      0.11316258 = sum of:
        0.11316258 = weight(_text_:system in 7403) [ClassicSimilarity], result of:
          0.11316258 = score(doc=7403,freq=16.0), product of:
            0.1642502 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.052150324 = queryNorm
            0.68896466 = fieldWeight in 7403, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0546875 = fieldNorm(doc=7403)
      0.2 = coord(1/5)
    
    Abstract
    Describes the TELMI monolingual text retrieval system and its future extension, a multilingual system. TELMI is designed for medium sized databases containing short texts. The characteristics of the system are fine-grained natural language processing (NLP); an open domain and a large scale knowledge base; automated indexing based on conceptual representation of texts and reusability of the NLP tools. Discusses the French MINITEL service, the MGS information service and the TELMI research system covering the full text system; NLP architecture; the lexical level; the syntactic level; the semantic level and an example of the use of a generic system
  12. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.02
    0.022610078 = product of:
      0.11305039 = sum of:
        0.11305039 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
          0.11305039 = score(doc=402,freq=2.0), product of:
            0.18262155 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052150324 = queryNorm
            0.61904186 = fieldWeight in 402, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.125 = fieldNorm(doc=402)
      0.2 = coord(1/5)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
  13. Fuhr, N.; Niewelt, B.: ¬Ein Retrievaltest mit automatisch indexierten Dokumenten (1984) 0.02
    0.019783817 = product of:
      0.098919086 = sum of:
        0.098919086 = weight(_text_:22 in 262) [ClassicSimilarity], result of:
          0.098919086 = score(doc=262,freq=2.0), product of:
            0.18262155 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052150324 = queryNorm
            0.5416616 = fieldWeight in 262, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.109375 = fieldNorm(doc=262)
      0.2 = coord(1/5)
    
    Date
    20.10.2000 12:22:23
  14. Hlava, M.M.K.: Automatic indexing : comparing rule-based and statistics-based indexing systems (2005) 0.02
    0.019783817 = product of:
      0.098919086 = sum of:
        0.098919086 = weight(_text_:22 in 6265) [ClassicSimilarity], result of:
          0.098919086 = score(doc=6265,freq=2.0), product of:
            0.18262155 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052150324 = queryNorm
            0.5416616 = fieldWeight in 6265, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.109375 = fieldNorm(doc=6265)
      0.2 = coord(1/5)
    
    Source
    Information outlook. 9(2005) no.8, S.22-23
  15. ¬The smart retrieval system : experiments in automatic document processing (1971) 0.02
    0.018289834 = product of:
      0.09144917 = sum of:
        0.09144917 = weight(_text_:system in 2330) [ClassicSimilarity], result of:
          0.09144917 = score(doc=2330,freq=2.0), product of:
            0.1642502 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.052150324 = queryNorm
            0.5567675 = fieldWeight in 2330, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.125 = fieldNorm(doc=2330)
      0.2 = coord(1/5)
    
  16. Dattola, R.T.: FIRST: Flexible information retrieval system for text (1979) 0.02
    0.018289834 = product of:
      0.09144917 = sum of:
        0.09144917 = weight(_text_:system in 5172) [ClassicSimilarity], result of:
          0.09144917 = score(doc=5172,freq=2.0), product of:
            0.1642502 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.052150324 = queryNorm
            0.5567675 = fieldWeight in 5172, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.125 = fieldNorm(doc=5172)
      0.2 = coord(1/5)
    
  17. Malone, L.C.; Driscoll, J.R.; Pepe, J.W.: Modeling the performance of an automated keywording system (1991) 0.02
    0.018289834 = product of:
      0.09144917 = sum of:
        0.09144917 = weight(_text_:system in 6682) [ClassicSimilarity], result of:
          0.09144917 = score(doc=6682,freq=8.0), product of:
            0.1642502 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.052150324 = queryNorm
            0.5567675 = fieldWeight in 6682, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.0625 = fieldNorm(doc=6682)
      0.2 = coord(1/5)
    
    Abstract
    Presents a model for predicting the performance of a computerised keyword assigning and indexing system. Statistical procedures were investigated in order to protect against incorrect keywording by the system behaving as an expert system designed to mimic the behaviour of human keyword indexers and representing lessons learned from military exercises and operations
  18. Malone, L.C.; Wildman-Pepe, J.; Driscoll, J.R.: Evaluation of an automated keywording system (1990) 0.02
    0.018146386 = product of:
      0.090731926 = sum of:
        0.090731926 = weight(_text_:system in 4999) [ClassicSimilarity], result of:
          0.090731926 = score(doc=4999,freq=14.0), product of:
            0.1642502 = queryWeight, product of:
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.052150324 = queryNorm
            0.5524007 = fieldWeight in 4999, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              3.1495528 = idf(docFreq=5152, maxDocs=44218)
              0.046875 = fieldNorm(doc=4999)
      0.2 = coord(1/5)
    
    Abstract
    An automated keywording system has been designed ro artifically behave as a human "expert" indexer. The system was designed to keyword 100 to 800 word documents representing lessons learned from military exercises and operations. A set of 74 documents can be keyworded on an IBM PS/2 model 80 in about five minutes. This paper presents a variety of ways for statistical documenting improvements in the development of an automated keywording system over time. It is not only beneficial to have some measure of system performance for a given time, but it is also useful as attemps are made to improve a system to assess if actual statistically significant improvements have been made. Furthermore, it is useful to identify the source of any existing problems so that they can be rectified. The specifics of the automated system that was evaluated are described, and the performance measures used are discussed.
  19. Fuhr, N.: Ranking-Experimente mit gewichteter Indexierung (1986) 0.02
    0.016957557 = product of:
      0.084787786 = sum of:
        0.084787786 = weight(_text_:22 in 58) [ClassicSimilarity], result of:
          0.084787786 = score(doc=58,freq=2.0), product of:
            0.18262155 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052150324 = queryNorm
            0.46428138 = fieldWeight in 58, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.09375 = fieldNorm(doc=58)
      0.2 = coord(1/5)
    
    Date
    14. 6.2015 22:12:44
  20. Hauer, M.: Automatische Indexierung (2000) 0.02
    0.016957557 = product of:
      0.084787786 = sum of:
        0.084787786 = weight(_text_:22 in 5887) [ClassicSimilarity], result of:
          0.084787786 = score(doc=5887,freq=2.0), product of:
            0.18262155 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052150324 = queryNorm
            0.46428138 = fieldWeight in 5887, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.09375 = fieldNorm(doc=5887)
      0.2 = coord(1/5)
    
    Source
    Wissen in Aktion: Wege des Knowledge Managements. 22. Online-Tagung der DGI, Frankfurt am Main, 2.-4.5.2000. Proceedings. Hrsg.: R. Schmidt

Years

Languages

Types

  • a 93
  • el 8
  • x 5
  • s 3
  • m 2
  • d 1
  • More… Less…