Search (93 results, page 1 of 5)

  • × theme_ss:"Automatisches Indexieren"
  1. Bordoni, L.; Pazienza, M.T.: Documents automatic indexing in an environmental domain (1997) 0.07
    0.06812288 = product of:
      0.13624576 = sum of:
        0.13624576 = sum of:
          0.087150216 = weight(_text_:language in 530) [ClassicSimilarity], result of:
            0.087150216 = score(doc=530,freq=4.0), product of:
              0.2030952 = queryWeight, product of:
                3.9232929 = idf(docFreq=2376, maxDocs=44218)
                0.051766515 = queryNorm
              0.42911017 = fieldWeight in 530, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.9232929 = idf(docFreq=2376, maxDocs=44218)
                0.0546875 = fieldNorm(doc=530)
          0.049095538 = weight(_text_:22 in 530) [ClassicSimilarity], result of:
            0.049095538 = score(doc=530,freq=2.0), product of:
              0.18127751 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051766515 = queryNorm
              0.2708308 = fieldWeight in 530, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=530)
      0.5 = coord(1/2)
    
    Abstract
    Describes an application of Natural Language Processing (NLP) techniques, in HIRMA (Hypertextual Information Retrieval Managed by ARIOSTO), to the problem of document indexing by referring to a system which incorporates natural language processing techniques to determine the subject of the text of documents and to associate them with relevant semantic indexes. Describes briefly the overall system, details of its implementation on a corpus of scientific abstracts related to environmental topics and experimental evidence of the system's behaviour. Analyzes in detail an experiment designed to evaluate the system's retrieval ability in terms of recall and precision
    Source
    International forum on information and documentation. 22(1997) no.1, S.17-28
  2. Hodges, P.R.: Keyword in title indexes : effectiveness of retrieval in computer searches (1983) 0.06
    0.05536002 = product of:
      0.11072004 = sum of:
        0.11072004 = sum of:
          0.0616245 = weight(_text_:language in 5001) [ClassicSimilarity], result of:
            0.0616245 = score(doc=5001,freq=2.0), product of:
              0.2030952 = queryWeight, product of:
                3.9232929 = idf(docFreq=2376, maxDocs=44218)
                0.051766515 = queryNorm
              0.30342668 = fieldWeight in 5001, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.9232929 = idf(docFreq=2376, maxDocs=44218)
                0.0546875 = fieldNorm(doc=5001)
          0.049095538 = weight(_text_:22 in 5001) [ClassicSimilarity], result of:
            0.049095538 = score(doc=5001,freq=2.0), product of:
              0.18127751 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051766515 = queryNorm
              0.2708308 = fieldWeight in 5001, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=5001)
      0.5 = coord(1/2)
    
    Abstract
    A study was done to test the effectiveness of retrieval using title word searching. It was based on actual search profiles used in the Mechanized Information Center at Ohio State University, in order ro replicate as closely as possible actual searching conditions. Fewer than 50% of the relevant titles were retrieved by keywords in titles. The low rate of retrieval can be attributes to three sources: titles themselves, user and information specialist ignorance of the subject vocabulary in use, and to general language problems. Across fields it was found that the social sciences had the best retrieval rate, with science having the next best, and arts and humanities the lowest. Ways to enhance and supplement keyword in title searching on the computer and in printed indexes are discussed.
    Date
    14. 3.1996 13:22:21
  3. Chowdhury, G.G.: Natural language processing and information retrieval : pt.1: basic issues; pt.2: major applications (1991) 0.04
    0.0440175 = product of:
      0.088035 = sum of:
        0.088035 = product of:
          0.17607 = sum of:
            0.17607 = weight(_text_:language in 3313) [ClassicSimilarity], result of:
              0.17607 = score(doc=3313,freq=8.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.8669334 = fieldWeight in 3313, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3313)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Reviews the basic issues and procedures involved in natural language processing of textual material for final use in information retrieval. Covers: natural language processing; natural language understanding; syntactic and semantic analysis; parsing; knowledge bases and knowledge representation
  4. Pritchard-Schoch, T.: Natural language comes of age (1993) 0.04
    0.039370455 = product of:
      0.07874091 = sum of:
        0.07874091 = product of:
          0.15748182 = sum of:
            0.15748182 = weight(_text_:language in 2570) [ClassicSimilarity], result of:
              0.15748182 = score(doc=2570,freq=10.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.77540886 = fieldWeight in 2570, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2570)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Discusses natural languages and the natural language implementations of Westlaw's full-text legal documents, Westlaw Is Natural. Natural language is not aritificial intelligence but a hybrid of linguistics, mathematics and statistics. Provides 3 classes of retrieval models. Explains how Westlaw processes an English query. Assesses WIN. Covers WIN enhancements; the natural language features of Congressional Quarterly's Washington Alert using a document for a query; the personal librarian front end search software and Dowquest from Dow Jones news/retrieval. Conmsiders whether natural language encourages fuzzy thinking and whether Boolean logic will still be needed
  5. Wolfram Language erkennt Bilder (2015) 0.04
    0.037350092 = product of:
      0.074700184 = sum of:
        0.074700184 = product of:
          0.14940037 = sum of:
            0.14940037 = weight(_text_:language in 1872) [ClassicSimilarity], result of:
              0.14940037 = score(doc=1872,freq=16.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.7356174 = fieldWeight in 1872, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1872)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Wolfram Research hat seine Cloud-basierte Programmiersprache Wolfram Language um eine Funktion zur Bilderkennung erweitert. Der Hersteller des Computeralgebrasystems Mathematica und Betreiber der Wissens-Suchmaschine Wolfram Alpha hat seinem System die Erkennung von Bildern beigebracht. Mit der Funktion ImageIdentify bekommt man in Wolfram Language jetzt zu einem Bild eine symbolische Beschreibung des Inhalts, die sich in der Sprache danach sogar weiterverarbeiten lässt. Als Demo dieser Funktion dient die Website The Wolfram Language Image Identification Project: Dort kann man ein beliebiges Bild hochladen und sich das Ergebnis anschauen. Die Website speichert einen Thumbnail des hochgeladenen Bildes, sodass man einen Link zu der Ergebnisseite weitergeben kann. Wie so oft bei künstlicher Intelligenz sind die Ergebnisse manchmal lustig daneben, oft aber auch überraschend gut. Die Funktion arbeitet mit einem neuronalen Netz, das mit einigen -zig Millionen Bildern trainiert wurde und etwa 10.000 Objekte identifizieren kann.
    Content
    Vgl.: http://www.imageidentify.com. Eine ausführlichere Erklärung der Funktionsweise und Hintergründe findet sich in Stephen Wolframs Blog-Eintrag: "Wolfram Language Artificial Intelligence: The Image Identification Project" unter: http://blog.stephenwolfram.com/2015/05/wolfram-language-artificial-intelligence-the-image-identification-project/. Vgl. auch: https://news.ycombinator.com/item?id=8621658.
    Object
    Wolfram Language
    Source
    http://www.heise.de/newsticker/meldung/Wolfram-Language-erkennt-Bilder-2650207.html?wt_mc=rss.ho.beitrag.rdf
  6. Jones, K.P.: Natural-language processing and automatic indexing : a reply (1990) 0.04
    0.035214003 = product of:
      0.07042801 = sum of:
        0.07042801 = product of:
          0.14085601 = sum of:
            0.14085601 = weight(_text_:language in 394) [ClassicSimilarity], result of:
              0.14085601 = score(doc=394,freq=2.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.6935467 = fieldWeight in 394, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.125 = fieldNorm(doc=394)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  7. Martins, A.L.; Souza, R.R.; Ribeiro de Mello, H.: ¬The use of noun phrases in information retrieval : proposing a mechanism for automatic classification (2014) 0.03
    0.031634297 = product of:
      0.063268594 = sum of:
        0.063268594 = sum of:
          0.035214003 = weight(_text_:language in 1441) [ClassicSimilarity], result of:
            0.035214003 = score(doc=1441,freq=2.0), product of:
              0.2030952 = queryWeight, product of:
                3.9232929 = idf(docFreq=2376, maxDocs=44218)
                0.051766515 = queryNorm
              0.17338668 = fieldWeight in 1441, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.9232929 = idf(docFreq=2376, maxDocs=44218)
                0.03125 = fieldNorm(doc=1441)
          0.028054593 = weight(_text_:22 in 1441) [ClassicSimilarity], result of:
            0.028054593 = score(doc=1441,freq=2.0), product of:
              0.18127751 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.051766515 = queryNorm
              0.15476047 = fieldWeight in 1441, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=1441)
      0.5 = coord(1/2)
    
    Abstract
    This paper presents a research on syntactic structures known as noun phrases (NP) being applied to increase the effectiveness and efficiency of the mechanisms for the document's classification. Our hypothesis is the fact that the NP can be used instead of single words as a semantic aggregator to reduce the number of words that will be used for the classification system without losing its semantic coverage, increasing its efficiency. The experiment divided the documents classification process in three phases: a) NP preprocessing b) system training; and c) classification experiments. In the first step, a corpus of digitalized texts was submitted to a natural language processing platform1 in which the part-of-speech tagging was done, and them PERL scripts pertaining to the PALAVRAS package were used to extract the Noun Phrases. The preprocessing also involved the tasks of a) removing NP low meaning pre-modifiers, as quantifiers; b) identification of synonyms and corresponding substitution for common hyperonyms; and c) stemming of the relevant words contained in the NP, for similitude checking with other NPs. The first tests with the resulting documents have demonstrated its effectiveness. We have compared the structural similarity of the documents before and after the whole pre-processing steps of phase one. The texts maintained the consistency with the original and have kept the readability. The second phase involves submitting the modified documents to a SVM algorithm to identify clusters and classify the documents. The classification rules are to be established using a machine learning approach. Finally, tests will be conducted to check the effectiveness of the whole process.
    Source
    Knowledge organization in the 21st century: between historical patterns and future prospects. Proceedings of the Thirteenth International ISKO Conference 19-22 May 2014, Kraków, Poland. Ed.: Wieslaw Babik
  8. Moens, M.F.: Automatic indexing and abstracting of document texts (2000) 0.03
    0.031125076 = product of:
      0.062250152 = sum of:
        0.062250152 = product of:
          0.124500304 = sum of:
            0.124500304 = weight(_text_:language in 6892) [ClassicSimilarity], result of:
              0.124500304 = score(doc=6892,freq=4.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.6130145 = fieldWeight in 6892, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.078125 = fieldNorm(doc=6892)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Content
    Need for indexing and abstracting texts; attributes of texts; text representations and their use; selection of natural language index terms; assignment of controlled language index texts; automatic abstracting; applications
  9. Kim, P.K.: ¬An automatic indexing of compound words based on mutual information for Korean text retrieval (1995) 0.03
    0.030496225 = product of:
      0.06099245 = sum of:
        0.06099245 = product of:
          0.1219849 = sum of:
            0.1219849 = weight(_text_:language in 620) [ClassicSimilarity], result of:
              0.1219849 = score(doc=620,freq=6.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.60062915 = fieldWeight in 620, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0625 = fieldNorm(doc=620)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Presents an automatic indexing technique for compound words suitable for an agglutinative language, specifically Korean. Discusses some construction conditions for compound words and the rules for decomposing compound words to enhance the exhaustivity of indexing, demonstrating that this system, mutual information, enhances both the exhaustivity of indexing and the specifity of terms. Suggests that the construction conditions and rules for decomposition presented may be used in multilingual information retrieval systems to translate the indexing terms of the specific language into those of the language required
  10. Vledutz-Stokolov, N.: Concept recognition in an automatic text-processing system for the life sciences (1987) 0.03
    0.029114842 = product of:
      0.058229685 = sum of:
        0.058229685 = product of:
          0.11645937 = sum of:
            0.11645937 = weight(_text_:language in 2849) [ClassicSimilarity], result of:
              0.11645937 = score(doc=2849,freq=14.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.57342255 = fieldWeight in 2849, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2849)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This article describes a natural-language text-processing system designed as an automatic aid to subject indexing at BIOSIS. The intellectual procedure the system should model is a deep indexing with a controlled vocabulary of biological concepts - Concept Headings (CHs). On the average, ten CHs are assigned to each article by BIOSIS indexers. The automatic procedure consists of two stages: (1) translation of natural-language biological titles into title-semantic representations which are in the constructed formalized language of Concept Primitives, and (2) translation of the latter representations into the language of CHs. The first stage is performed by matching the titles agianst the system's Semantic Vocabulary (SV). The SV currently contains approximately 15.000 biological natural-language terms and their translations in the language of Concept Primitives. Tor the ambiguous terms, the SV contains the algorithmical rules of term disambiguation, ruels based on semantic analysis of the contexts. The second stage of the automatic procedure is performed by matching the title representations against the CH definitions, formulated as Boolean search strategies in the language of Concept Primitives. Three experiments performed with the system and their results are decribed. The most typical problems the system encounters, the problems of lexical and situational ambiguities, are discussed. The disambiguation techniques employed are described and demonstrated in many examples
  11. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.03
    0.028054593 = product of:
      0.056109186 = sum of:
        0.056109186 = product of:
          0.11221837 = sum of:
            0.11221837 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
              0.11221837 = score(doc=402,freq=2.0), product of:
                0.18127751 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051766515 = queryNorm
                0.61904186 = fieldWeight in 402, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=402)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
  12. Cohen, J.D.: Highlights: language- and domain-independent automatic indexing terms for abstracting (1995) 0.03
    0.026684195 = product of:
      0.05336839 = sum of:
        0.05336839 = product of:
          0.10673678 = sum of:
            0.10673678 = weight(_text_:language in 1793) [ClassicSimilarity], result of:
              0.10673678 = score(doc=1793,freq=6.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.5255505 = fieldWeight in 1793, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1793)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Presents a model of drawing index terms from text. The approach uses no stop list, stemmer, or other language and domain specific component, allowing operation in any language or domain with only trivial modification. The method uses n-grams counts, achieving a function similar to, but more general than, a stemmer. The generated index terms, called 'highlights', are suitable for identifying the topic for perusal and selection. An extension is also described and demonstrated which selects index terms to represent a subset of documents, distinguishing them from the corpus. Presents some experimental results, showing operation in English, Spanish, German, Georgian, Russian and Japanese
  13. Taylor, S.L.: Integrating natural language understanding with document structure analysis (1994) 0.03
    0.026684195 = product of:
      0.05336839 = sum of:
        0.05336839 = product of:
          0.10673678 = sum of:
            0.10673678 = weight(_text_:language in 1794) [ClassicSimilarity], result of:
              0.10673678 = score(doc=1794,freq=6.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.5255505 = fieldWeight in 1794, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1794)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Document understanding, the interpretation of a document from its image form, is a technology area which benefits greatly from the integration of natural language processing with image processing. Develops a prototype of an Intelligent Document Understanding System (IDUS) which employs several technologies: image processing, optical character recognition, document structure analysis and text understanding in a cooperative fashion. Discusses those areas of research during development of IDUS where it is found that the most benefit from the integration of natural language processing and image processing occured: document structure analysis, OCR correction, and text analysis. Discusses 2 applications which are supported by IDUS: text retrieval and automatic generation of hypertext links
  14. Silvester, J.P.; Genuardi, M.T.: Machine-aided indexing from the analysis of natural language text (1994) 0.03
    0.0264105 = product of:
      0.052821 = sum of:
        0.052821 = product of:
          0.105642 = sum of:
            0.105642 = weight(_text_:language in 2989) [ClassicSimilarity], result of:
              0.105642 = score(doc=2989,freq=2.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.52016 = fieldWeight in 2989, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.09375 = fieldNorm(doc=2989)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  15. Kiros, R.; Salakhutdinov, R.; Zemel, R.S.: Unifying visual-semantic embeddings with multimodal neural language models (2014) 0.03
    0.0264105 = product of:
      0.052821 = sum of:
        0.052821 = product of:
          0.105642 = sum of:
            0.105642 = weight(_text_:language in 1871) [ClassicSimilarity], result of:
              0.105642 = score(doc=1871,freq=8.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.52016 = fieldWeight in 1871, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1871)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Inspired by recent advances in multimodal learning and machine translation, we introduce an encoder-decoder pipeline that learns (a): a multimodal joint embedding space with images and text and (b): a novel language model for decoding distributed representations from our space. Our pipeline effectively unifies joint image-text embedding models with multimodal neural language models. We introduce the structure-content neural language model that disentangles the structure of a sentence to its content, conditioned on representations produced by the encoder. The encoder allows one to rank images and sentences while the decoder can generate novel descriptions from scratch. Using LSTM to encode sentences, we match the state-of-the-art performance on Flickr8K and Flickr30K without using object detections. We also set new best results when using the 19-layer Oxford convolutional network. Furthermore we show that with linear encoders, the learned embedding space captures multimodal regularities in terms of vector space arithmetic e.g. *image of a blue car* - "blue" + "red" is near images of red cars. Sample captions generated for 800 images are made available for comparison.
  16. Koryconski, C.; Newell, A.F.: Natural-language processing and automatic indexing (1990) 0.02
    0.02490006 = product of:
      0.04980012 = sum of:
        0.04980012 = product of:
          0.09960024 = sum of:
            0.09960024 = weight(_text_:language in 2313) [ClassicSimilarity], result of:
              0.09960024 = score(doc=2313,freq=4.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.4904116 = fieldWeight in 2313, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2313)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The task of producing satisfactory indexes by automatic means has been tackled on two fronts: by statistical analysis of text and by attempting content analysis of the text in much the same way as a human indexer does. Though statistical techniques have a lot to offer for free-text database systems, neither method has had much success with back-of-the-book indexing. This review examines some problems associated with the application of natural-language processing techniques to book texts. - Vgl. auch die Erwiderung von K.P. Jones
  17. Smart, G.: Using language analysis to manage information (1993) 0.02
    0.02490006 = product of:
      0.04980012 = sum of:
        0.04980012 = product of:
          0.09960024 = sum of:
            0.09960024 = weight(_text_:language in 4423) [ClassicSimilarity], result of:
              0.09960024 = score(doc=4423,freq=4.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.4904116 = fieldWeight in 4423, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4423)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    The ESPRIT project SIMPR developed software to analyse documents and generate indexes for them. Of immediate application as a document indexing and classification system, this also offers a technology for information modelling that has broader implications, supporting many new uses for information management softeware. The project was based on the assumption that information can only be managed successfully by computer systems that can view the information contained in a document through the language in which the document is written, and that systems need to be sufficiently flexible to respond to the changing requirements of document use
  18. Samstag-Schnock, U.; Meadow, C.T.: PBS: an ecomical natural language query interpreter (1993) 0.02
    0.02490006 = product of:
      0.04980012 = sum of:
        0.04980012 = product of:
          0.09960024 = sum of:
            0.09960024 = weight(_text_:language in 5091) [ClassicSimilarity], result of:
              0.09960024 = score(doc=5091,freq=4.0), product of:
                0.2030952 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.051766515 = queryNorm
                0.4904116 = fieldWeight in 5091, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5091)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Reports on the design and implementation of the information searching and retrieval software, PBS (Parsing, Boolean recognition, Stemming) for the front end OAK 2, a new version of OAK developed at Toronto Univ. OAK 2 is a research tool for user behaviour studies. PBS receives natural language search statements from an end user and identifies search facets and implied Boolean logic operators
  19. Fuhr, N.; Niewelt, B.: ¬Ein Retrievaltest mit automatisch indexierten Dokumenten (1984) 0.02
    0.024547769 = product of:
      0.049095538 = sum of:
        0.049095538 = product of:
          0.098191075 = sum of:
            0.098191075 = weight(_text_:22 in 262) [ClassicSimilarity], result of:
              0.098191075 = score(doc=262,freq=2.0), product of:
                0.18127751 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051766515 = queryNorm
                0.5416616 = fieldWeight in 262, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=262)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    20.10.2000 12:22:23
  20. Hlava, M.M.K.: Automatic indexing : comparing rule-based and statistics-based indexing systems (2005) 0.02
    0.024547769 = product of:
      0.049095538 = sum of:
        0.049095538 = product of:
          0.098191075 = sum of:
            0.098191075 = weight(_text_:22 in 6265) [ClassicSimilarity], result of:
              0.098191075 = score(doc=6265,freq=2.0), product of:
                0.18127751 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.051766515 = queryNorm
                0.5416616 = fieldWeight in 6265, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=6265)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Information outlook. 9(2005) no.8, S.22-23

Years

Languages

  • e 67
  • d 21
  • f 2
  • ja 1
  • ru 1
  • sp 1
  • More… Less…

Types

  • a 80
  • el 12
  • m 2
  • s 2
  • x 2
  • d 1
  • More… Less…