Search (117 results, page 1 of 6)

  • × theme_ss:"Automatisches Indexieren"
  1. Kim, P.K.: ¬An automatic indexing of compound words based on mutual information for Korean text retrieval (1995) 0.02
    0.024179649 = product of:
      0.096718594 = sum of:
        0.096718594 = product of:
          0.14507788 = sum of:
            0.09908719 = weight(_text_:language in 620) [ClassicSimilarity], result of:
              0.09908719 = score(doc=620,freq=6.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.60062915 = fieldWeight in 620, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0625 = fieldNorm(doc=620)
            0.045990683 = weight(_text_:29 in 620) [ClassicSimilarity], result of:
              0.045990683 = score(doc=620,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.31092256 = fieldWeight in 620, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=620)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Abstract
    Presents an automatic indexing technique for compound words suitable for an agglutinative language, specifically Korean. Discusses some construction conditions for compound words and the rules for decomposing compound words to enhance the exhaustivity of indexing, demonstrating that this system, mutual information, enhances both the exhaustivity of indexing and the specifity of terms. Suggests that the construction conditions and rules for decomposition presented may be used in multilingual information retrieval systems to translate the indexing terms of the specific language into those of the language required
    Source
    Library and information science. 1995, no.34, S.29-38
  2. Koryconski, C.; Newell, A.F.: Natural-language processing and automatic indexing (1990) 0.02
    0.021149173 = product of:
      0.08459669 = sum of:
        0.08459669 = product of:
          0.12689504 = sum of:
            0.08090435 = weight(_text_:language in 2313) [ClassicSimilarity], result of:
              0.08090435 = score(doc=2313,freq=4.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.4904116 = fieldWeight in 2313, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2313)
            0.045990683 = weight(_text_:29 in 2313) [ClassicSimilarity], result of:
              0.045990683 = score(doc=2313,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.31092256 = fieldWeight in 2313, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2313)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Abstract
    The task of producing satisfactory indexes by automatic means has been tackled on two fronts: by statistical analysis of text and by attempting content analysis of the text in much the same way as a human indexer does. Though statistical techniques have a lot to offer for free-text database systems, neither method has had much success with back-of-the-book indexing. This review examines some problems associated with the application of natural-language processing techniques to book texts. - Vgl. auch die Erwiderung von K.P. Jones
    Source
    Indexer. 17(1990), S.21-29
  3. Bordoni, L.; Pazienza, M.T.: Documents automatic indexing in an environmental domain (1997) 0.02
    0.018445192 = product of:
      0.07378077 = sum of:
        0.07378077 = product of:
          0.11067115 = sum of:
            0.070791304 = weight(_text_:language in 530) [ClassicSimilarity], result of:
              0.070791304 = score(doc=530,freq=4.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.42911017 = fieldWeight in 530, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=530)
            0.039879844 = weight(_text_:22 in 530) [ClassicSimilarity], result of:
              0.039879844 = score(doc=530,freq=2.0), product of:
                0.14725003 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.042049456 = queryNorm
                0.2708308 = fieldWeight in 530, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=530)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Abstract
    Describes an application of Natural Language Processing (NLP) techniques, in HIRMA (Hypertextual Information Retrieval Managed by ARIOSTO), to the problem of document indexing by referring to a system which incorporates natural language processing techniques to determine the subject of the text of documents and to associate them with relevant semantic indexes. Describes briefly the overall system, details of its implementation on a corpus of scientific abstracts related to environmental topics and experimental evidence of the system's behaviour. Analyzes in detail an experiment designed to evaluate the system's retrieval ability in terms of recall and precision
    Source
    International forum on information and documentation. 22(1997) no.1, S.17-28
  4. Damerau, F.J.: Generating an evaluating domain-oriented multi-word terms from texts (1993) 0.02
    0.017199783 = product of:
      0.06879913 = sum of:
        0.06879913 = product of:
          0.10319869 = sum of:
            0.05720801 = weight(_text_:language in 5814) [ClassicSimilarity], result of:
              0.05720801 = score(doc=5814,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.34677336 = fieldWeight in 5814, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5814)
            0.045990683 = weight(_text_:29 in 5814) [ClassicSimilarity], result of:
              0.045990683 = score(doc=5814,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.31092256 = fieldWeight in 5814, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5814)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Abstract
    Examines techniques for automatically generating domain vocabularies from large text collections. Focuses on the problem of generating multi-word vocabulary terms (specifically pairs). Discusses statistical issues associated with word co-occurrences likely to be of use in a natural language interface. Provides a more objective evaluation of the selection procedures. As substantial experimentation with subjects using a working query system is absent, all evaluation is necessarily subjective. Uses surrogate for experimentation by relying on pre-existing dictionaries as indicators of domain relevance
    Source
    Information processing and management. 29(1993) no.4, S.433-447
  5. Schulz, K.U.; Brunner, L.: Vollautomatische thematische Verschlagwortung großer Textkollektionen mittels semantischer Netze (2017) 0.02
    0.01504981 = product of:
      0.06019924 = sum of:
        0.06019924 = product of:
          0.090298854 = sum of:
            0.05005701 = weight(_text_:language in 3493) [ClassicSimilarity], result of:
              0.05005701 = score(doc=3493,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.30342668 = fieldWeight in 3493, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3493)
            0.040241845 = weight(_text_:29 in 3493) [ClassicSimilarity], result of:
              0.040241845 = score(doc=3493,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.27205724 = fieldWeight in 3493, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3493)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Source
    Theorie, Semantik und Organisation von Wissen: Proceedings der 13. Tagung der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) und dem 13. Internationalen Symposium der Informationswissenschaft der Higher Education Association for Information Science (HI) Potsdam (19.-20.03.2013): 'Theory, Information and Organization of Knowledge' / Proceedings der 14. Tagung der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) und Natural Language & Information Systems (NLDB) Passau (16.06.2015): 'Lexical Resources for Knowledge Organization' / Proceedings des Workshops der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) auf der SEMANTICS Leipzig (1.09.2014): 'Knowledge Organization and Semantic Web' / Proceedings des Workshops der Polnischen und Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) Cottbus (29.-30.09.2011): 'Economics of Knowledge Production and Organization'. Hrsg. von W. Babik, H.P. Ohly u. K. Weber
  6. Böhm, A.; Seifert, C.; Schlötterer, J.; Granitzer, M.: Identifying tweets from the economic domain (2017) 0.02
    0.01504981 = product of:
      0.06019924 = sum of:
        0.06019924 = product of:
          0.090298854 = sum of:
            0.05005701 = weight(_text_:language in 3495) [ClassicSimilarity], result of:
              0.05005701 = score(doc=3495,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.30342668 = fieldWeight in 3495, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3495)
            0.040241845 = weight(_text_:29 in 3495) [ClassicSimilarity], result of:
              0.040241845 = score(doc=3495,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.27205724 = fieldWeight in 3495, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3495)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Source
    Theorie, Semantik und Organisation von Wissen: Proceedings der 13. Tagung der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) und dem 13. Internationalen Symposium der Informationswissenschaft der Higher Education Association for Information Science (HI) Potsdam (19.-20.03.2013): 'Theory, Information and Organization of Knowledge' / Proceedings der 14. Tagung der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) und Natural Language & Information Systems (NLDB) Passau (16.06.2015): 'Lexical Resources for Knowledge Organization' / Proceedings des Workshops der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) auf der SEMANTICS Leipzig (1.09.2014): 'Knowledge Organization and Semantic Web' / Proceedings des Workshops der Polnischen und Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) Cottbus (29.-30.09.2011): 'Economics of Knowledge Production and Organization'. Hrsg. von W. Babik, H.P. Ohly u. K. Weber
  7. Kempf, A.O.: Neue Verfahrenswege der Wissensorganisation : eine Evaluation automatischer Indexierung in der sozialwissenschaftlichen Fachinformation (2017) 0.02
    0.01504981 = product of:
      0.06019924 = sum of:
        0.06019924 = product of:
          0.090298854 = sum of:
            0.05005701 = weight(_text_:language in 3497) [ClassicSimilarity], result of:
              0.05005701 = score(doc=3497,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.30342668 = fieldWeight in 3497, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3497)
            0.040241845 = weight(_text_:29 in 3497) [ClassicSimilarity], result of:
              0.040241845 = score(doc=3497,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.27205724 = fieldWeight in 3497, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3497)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Source
    Theorie, Semantik und Organisation von Wissen: Proceedings der 13. Tagung der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) und dem 13. Internationalen Symposium der Informationswissenschaft der Higher Education Association for Information Science (HI) Potsdam (19.-20.03.2013): 'Theory, Information and Organization of Knowledge' / Proceedings der 14. Tagung der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) und Natural Language & Information Systems (NLDB) Passau (16.06.2015): 'Lexical Resources for Knowledge Organization' / Proceedings des Workshops der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) auf der SEMANTICS Leipzig (1.09.2014): 'Knowledge Organization and Semantic Web' / Proceedings des Workshops der Polnischen und Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation (ISKO) Cottbus (29.-30.09.2011): 'Economics of Knowledge Production and Organization'. Hrsg. von W. Babik, H.P. Ohly u. K. Weber
  8. Hodges, P.R.: Keyword in title indexes : effectiveness of retrieval in computer searches (1983) 0.01
    0.014989476 = product of:
      0.059957903 = sum of:
        0.059957903 = product of:
          0.08993685 = sum of:
            0.05005701 = weight(_text_:language in 5001) [ClassicSimilarity], result of:
              0.05005701 = score(doc=5001,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.30342668 = fieldWeight in 5001, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5001)
            0.039879844 = weight(_text_:22 in 5001) [ClassicSimilarity], result of:
              0.039879844 = score(doc=5001,freq=2.0), product of:
                0.14725003 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.042049456 = queryNorm
                0.2708308 = fieldWeight in 5001, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5001)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Abstract
    A study was done to test the effectiveness of retrieval using title word searching. It was based on actual search profiles used in the Mechanized Information Center at Ohio State University, in order ro replicate as closely as possible actual searching conditions. Fewer than 50% of the relevant titles were retrieved by keywords in titles. The low rate of retrieval can be attributes to three sources: titles themselves, user and information specialist ignorance of the subject vocabulary in use, and to general language problems. Across fields it was found that the social sciences had the best retrieval rate, with science having the next best, and arts and humanities the lowest. Ways to enhance and supplement keyword in title searching on the computer and in printed indexes are discussed.
    Date
    14. 3.1996 13:22:21
  9. Wolfekuhler, M.R.; Punch, W.F.: Finding salient features for personal Web pages categories (1997) 0.01
    0.013353615 = product of:
      0.05341446 = sum of:
        0.05341446 = product of:
          0.08012169 = sum of:
            0.040241845 = weight(_text_:29 in 2673) [ClassicSimilarity], result of:
              0.040241845 = score(doc=2673,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.27205724 = fieldWeight in 2673, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2673)
            0.039879844 = weight(_text_:22 in 2673) [ClassicSimilarity], result of:
              0.039879844 = score(doc=2673,freq=2.0), product of:
                0.14725003 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.042049456 = queryNorm
                0.2708308 = fieldWeight in 2673, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2673)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Date
    1. 8.1996 22:08:06
    Source
    Computer networks and ISDN systems. 29(1997) no.8, S.1147-1156
  10. Franke-Maier, M.: Anforderungen an die Qualität der Inhaltserschließung im Spannungsfeld von intellektuell und automatisch erzeugten Metadaten (2018) 0.01
    0.013353615 = product of:
      0.05341446 = sum of:
        0.05341446 = product of:
          0.08012169 = sum of:
            0.040241845 = weight(_text_:29 in 5344) [ClassicSimilarity], result of:
              0.040241845 = score(doc=5344,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.27205724 = fieldWeight in 5344, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5344)
            0.039879844 = weight(_text_:22 in 5344) [ClassicSimilarity], result of:
              0.039879844 = score(doc=5344,freq=2.0), product of:
                0.14725003 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.042049456 = queryNorm
                0.2708308 = fieldWeight in 5344, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5344)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Abstract
    Spätestens seit dem Deutschen Bibliothekartag 2018 hat sich die Diskussion zu den automatischen Verfahren der Inhaltserschließung der Deutschen Nationalbibliothek von einer politisch geführten Diskussion in eine Qualitätsdiskussion verwandelt. Der folgende Beitrag beschäftigt sich mit Fragen der Qualität von Inhaltserschließung in digitalen Zeiten, wo heterogene Erzeugnisse unterschiedlicher Verfahren aufeinandertreffen und versucht, wichtige Anforderungen an Qualität zu definieren. Dieser Tagungsbeitrag fasst die vom Autor als Impulse vorgetragenen Ideen beim Workshop der FAG "Erschließung und Informationsvermittlung" des GBV am 29. August 2018 in Kiel zusammen. Der Workshop fand im Rahmen der 22. Verbundkonferenz des GBV statt.
  11. Li, W.; Wong, K.-F.; Yuan, C.: Toward automatic Chinese temporal information extraction (2001) 0.01
    0.013218233 = product of:
      0.052872933 = sum of:
        0.052872933 = product of:
          0.0793094 = sum of:
            0.050565217 = weight(_text_:language in 6029) [ClassicSimilarity], result of:
              0.050565217 = score(doc=6029,freq=4.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.30650726 = fieldWeight in 6029, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=6029)
            0.028744178 = weight(_text_:29 in 6029) [ClassicSimilarity], result of:
              0.028744178 = score(doc=6029,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.19432661 = fieldWeight in 6029, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=6029)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Abstract
    Over the past few years, temporal information processing and temporal database management have increasingly become hot topics. Nevertheless, only a few researchers have investigated these areas in the Chinese language. This lays down the objective of our research: to exploit Chinese language processing techniques for temporal information extraction and concept reasoning. In this article, we first study the mechanism for expressing time in Chinese. On the basis of the study, we then design a general frame structure for maintaining the extracted temporal concepts and propose a system for extracting time-dependent information from Hong Kong financial news. In the system, temporal knowledge is represented by different types of temporal concepts (TTC) and different temporal relations, including absolute and relative relations, which are used to correlate between action times and reference times. In analyzing a sentence, the algorithm first determines the situation related to the verb. This in turn will identify the type of temporal concept associated with the verb. After that, the relevant temporal information is extracted and the temporal relations are derived. These relations link relevant concept frames together in chronological order, which in turn provide the knowledge to fulfill users' queries, e.g., for question-answering (i.e., Q&A) applications
    Date
    29. 9.2001 14:02:50
  12. Chowdhury, G.G.: Natural language processing and information retrieval : pt.1: basic issues; pt.2: major applications (1991) 0.01
    0.011918336 = product of:
      0.047673345 = sum of:
        0.047673345 = product of:
          0.14302003 = sum of:
            0.14302003 = weight(_text_:language in 3313) [ClassicSimilarity], result of:
              0.14302003 = score(doc=3313,freq=8.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.8669334 = fieldWeight in 3313, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3313)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    Reviews the basic issues and procedures involved in natural language processing of textual material for final use in information retrieval. Covers: natural language processing; natural language understanding; syntactic and semantic analysis; parsing; knowledge bases and knowledge representation
  13. Pritchard-Schoch, T.: Natural language comes of age (1993) 0.01
    0.010660085 = product of:
      0.04264034 = sum of:
        0.04264034 = product of:
          0.12792102 = sum of:
            0.12792102 = weight(_text_:language in 2570) [ClassicSimilarity], result of:
              0.12792102 = score(doc=2570,freq=10.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.77540886 = fieldWeight in 2570, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2570)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    Discusses natural languages and the natural language implementations of Westlaw's full-text legal documents, Westlaw Is Natural. Natural language is not aritificial intelligence but a hybrid of linguistics, mathematics and statistics. Provides 3 classes of retrieval models. Explains how Westlaw processes an English query. Assesses WIN. Covers WIN enhancements; the natural language features of Congressional Quarterly's Washington Alert using a document for a query; the personal librarian front end search software and Dowquest from Dow Jones news/retrieval. Conmsiders whether natural language encourages fuzzy thinking and whether Boolean logic will still be needed
  14. Wolfram Language erkennt Bilder (2015) 0.01
    0.010113044 = product of:
      0.040452175 = sum of:
        0.040452175 = product of:
          0.12135652 = sum of:
            0.12135652 = weight(_text_:language in 1872) [ClassicSimilarity], result of:
              0.12135652 = score(doc=1872,freq=16.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.7356174 = fieldWeight in 1872, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1872)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    Wolfram Research hat seine Cloud-basierte Programmiersprache Wolfram Language um eine Funktion zur Bilderkennung erweitert. Der Hersteller des Computeralgebrasystems Mathematica und Betreiber der Wissens-Suchmaschine Wolfram Alpha hat seinem System die Erkennung von Bildern beigebracht. Mit der Funktion ImageIdentify bekommt man in Wolfram Language jetzt zu einem Bild eine symbolische Beschreibung des Inhalts, die sich in der Sprache danach sogar weiterverarbeiten lässt. Als Demo dieser Funktion dient die Website The Wolfram Language Image Identification Project: Dort kann man ein beliebiges Bild hochladen und sich das Ergebnis anschauen. Die Website speichert einen Thumbnail des hochgeladenen Bildes, sodass man einen Link zu der Ergebnisseite weitergeben kann. Wie so oft bei künstlicher Intelligenz sind die Ergebnisse manchmal lustig daneben, oft aber auch überraschend gut. Die Funktion arbeitet mit einem neuronalen Netz, das mit einigen -zig Millionen Bildern trainiert wurde und etwa 10.000 Objekte identifizieren kann.
    Content
    Vgl.: http://www.imageidentify.com. Eine ausführlichere Erklärung der Funktionsweise und Hintergründe findet sich in Stephen Wolframs Blog-Eintrag: "Wolfram Language Artificial Intelligence: The Image Identification Project" unter: http://blog.stephenwolfram.com/2015/05/wolfram-language-artificial-intelligence-the-image-identification-project/. Vgl. auch: https://news.ycombinator.com/item?id=8621658.
    Object
    Wolfram Language
    Source
    http://www.heise.de/newsticker/meldung/Wolfram-Language-erkennt-Bilder-2650207.html?wt_mc=rss.ho.beitrag.rdf
  15. Jones, K.P.: Natural-language processing and automatic indexing : a reply (1990) 0.01
    0.009534668 = product of:
      0.038138673 = sum of:
        0.038138673 = product of:
          0.11441602 = sum of:
            0.11441602 = weight(_text_:language in 394) [ClassicSimilarity], result of:
              0.11441602 = score(doc=394,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.6935467 = fieldWeight in 394, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.125 = fieldNorm(doc=394)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
  16. Kuhlen, R.: Morphologische Relationen durch Reduktionsalgorithmen (1974) 0.01
    0.009485095 = product of:
      0.03794038 = sum of:
        0.03794038 = product of:
          0.113821134 = sum of:
            0.113821134 = weight(_text_:29 in 4251) [ClassicSimilarity], result of:
              0.113821134 = score(doc=4251,freq=4.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.7694941 = fieldWeight in 4251, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4251)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    29. 1.2011 14:56:29
  17. Martins, A.L.; Souza, R.R.; Ribeiro de Mello, H.: ¬The use of noun phrases in information retrieval : proposing a mechanism for automatic classification (2014) 0.01
    0.008565415 = product of:
      0.03426166 = sum of:
        0.03426166 = product of:
          0.05139249 = sum of:
            0.028604005 = weight(_text_:language in 1441) [ClassicSimilarity], result of:
              0.028604005 = score(doc=1441,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.17338668 = fieldWeight in 1441, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1441)
            0.022788482 = weight(_text_:22 in 1441) [ClassicSimilarity], result of:
              0.022788482 = score(doc=1441,freq=2.0), product of:
                0.14725003 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.042049456 = queryNorm
                0.15476047 = fieldWeight in 1441, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1441)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Abstract
    This paper presents a research on syntactic structures known as noun phrases (NP) being applied to increase the effectiveness and efficiency of the mechanisms for the document's classification. Our hypothesis is the fact that the NP can be used instead of single words as a semantic aggregator to reduce the number of words that will be used for the classification system without losing its semantic coverage, increasing its efficiency. The experiment divided the documents classification process in three phases: a) NP preprocessing b) system training; and c) classification experiments. In the first step, a corpus of digitalized texts was submitted to a natural language processing platform1 in which the part-of-speech tagging was done, and them PERL scripts pertaining to the PALAVRAS package were used to extract the Noun Phrases. The preprocessing also involved the tasks of a) removing NP low meaning pre-modifiers, as quantifiers; b) identification of synonyms and corresponding substitution for common hyperonyms; and c) stemming of the relevant words contained in the NP, for similitude checking with other NPs. The first tests with the resulting documents have demonstrated its effectiveness. We have compared the structural similarity of the documents before and after the whole pre-processing steps of phase one. The texts maintained the consistency with the original and have kept the readability. The second phase involves submitting the modified documents to a SVM algorithm to identify clusters and classify the documents. The classification rules are to be established using a machine learning approach. Finally, tests will be conducted to check the effectiveness of the whole process.
    Source
    Knowledge organization in the 21st century: between historical patterns and future prospects. Proceedings of the Thirteenth International ISKO Conference 19-22 May 2014, Kraków, Poland. Ed.: Wieslaw Babik
  18. Moens, M.F.: Automatic indexing and abstracting of document texts (2000) 0.01
    0.008427536 = product of:
      0.033710144 = sum of:
        0.033710144 = product of:
          0.10113043 = sum of:
            0.10113043 = weight(_text_:language in 6892) [ClassicSimilarity], result of:
              0.10113043 = score(doc=6892,freq=4.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.6130145 = fieldWeight in 6892, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.078125 = fieldNorm(doc=6892)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Content
    Need for indexing and abstracting texts; attributes of texts; text representations and their use; selection of natural language index terms; assignment of controlled language index texts; automatic abstracting; applications
  19. Vledutz-Stokolov, N.: Concept recognition in an automatic text-processing system for the life sciences (1987) 0.01
    0.007883239 = product of:
      0.031532954 = sum of:
        0.031532954 = product of:
          0.09459886 = sum of:
            0.09459886 = weight(_text_:language in 2849) [ClassicSimilarity], result of:
              0.09459886 = score(doc=2849,freq=14.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.57342255 = fieldWeight in 2849, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2849)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    This article describes a natural-language text-processing system designed as an automatic aid to subject indexing at BIOSIS. The intellectual procedure the system should model is a deep indexing with a controlled vocabulary of biological concepts - Concept Headings (CHs). On the average, ten CHs are assigned to each article by BIOSIS indexers. The automatic procedure consists of two stages: (1) translation of natural-language biological titles into title-semantic representations which are in the constructed formalized language of Concept Primitives, and (2) translation of the latter representations into the language of CHs. The first stage is performed by matching the titles agianst the system's Semantic Vocabulary (SV). The SV currently contains approximately 15.000 biological natural-language terms and their translations in the language of Concept Primitives. Tor the ambiguous terms, the SV contains the algorithmical rules of term disambiguation, ruels based on semantic analysis of the contexts. The second stage of the automatic procedure is performed by matching the title representations against the CH definitions, formulated as Boolean search strategies in the language of Concept Primitives. Three experiments performed with the system and their results are decribed. The most typical problems the system encounters, the problems of lexical and situational ambiguities, are discussed. The disambiguation techniques employed are described and demonstrated in many examples
  20. Panyr, J.: STEINADLER: ein Verfahren zur automatischen Deskribierung und zur automatischen thematischen Klassifikation (1978) 0.01
    0.007665114 = product of:
      0.030660456 = sum of:
        0.030660456 = product of:
          0.09198137 = sum of:
            0.09198137 = weight(_text_:29 in 5169) [ClassicSimilarity], result of:
              0.09198137 = score(doc=5169,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.6218451 = fieldWeight in 5169, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.125 = fieldNorm(doc=5169)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Source
    Nachrichten für Dokumentation. 29(1978), S.92-96

Years

Languages

Types

  • a 101
  • el 14
  • m 3
  • x 3
  • s 2
  • d 1
  • p 1
  • More… Less…