Search (43 results, page 1 of 3)

  • × theme_ss:"Automatisches Indexieren"
  • × theme_ss:"Computerlinguistik"
  1. Werner, H.: Indexierung auf linguistischer Grundlage am Beispiel von JUDO-DS(1) (1982) 0.02
    0.020341193 = product of:
      0.040682387 = sum of:
        0.040682387 = product of:
          0.061023578 = sum of:
            0.008065818 = weight(_text_:a in 3017) [ClassicSimilarity], result of:
              0.008065818 = score(doc=3017,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.15287387 = fieldWeight in 3017, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3017)
            0.052957762 = weight(_text_:h in 3017) [ClassicSimilarity], result of:
              0.052957762 = score(doc=3017,freq=4.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.4658342 = fieldWeight in 3017, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3017)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Deutscher Dokumentartag 1981, Mainz, 5.-8.10.1981: Kleincomputer in Information und Dokumentation. Bearb.: H. Strohl-Goebel
    Type
    a
  2. Riloff, E.: ¬An empirical study of automated dictionary construction for information extraction in three domains (1996) 0.02
    0.01906709 = product of:
      0.03813418 = sum of:
        0.03813418 = product of:
          0.05720127 = sum of:
            0.007604526 = weight(_text_:a in 6752) [ClassicSimilarity], result of:
              0.007604526 = score(doc=6752,freq=4.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.14413087 = fieldWeight in 6752, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6752)
            0.049596746 = weight(_text_:22 in 6752) [ClassicSimilarity], result of:
              0.049596746 = score(doc=6752,freq=2.0), product of:
                0.16023713 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045758117 = queryNorm
                0.30952093 = fieldWeight in 6752, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6752)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Abstract
    AutoSlog is a system that addresses the knowledge engineering bottleneck for information extraction. AutoSlog automatically creates domain specific dictionaries for information extraction, given an appropriate training corpus. Describes experiments with AutoSlog in terrorism, joint ventures and microelectronics domains. Compares the performance of AutoSlog across the 3 domains, discusses the lessons learned and presents results from 2 experiments which demonstrate that novice users can generate effective dictionaries using AutoSlog
    Date
    6. 3.1997 16:22:15
    Type
    a
  3. Hoppe, A.: ¬Die systematischen Grundlagen für ein linguistisch orientiertes maschinelles Dokumentationsverfahren (1969) 0.02
    0.018998615 = product of:
      0.03799723 = sum of:
        0.03799723 = product of:
          0.056995846 = sum of:
            0.013307921 = weight(_text_:a in 4720) [ClassicSimilarity], result of:
              0.013307921 = score(doc=4720,freq=4.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.25222903 = fieldWeight in 4720, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4720)
            0.043687925 = weight(_text_:h in 4720) [ClassicSimilarity], result of:
              0.043687925 = score(doc=4720,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.38429362 = fieldWeight in 4720, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4720)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Nachrichten für Dokumentation. 20(1969) H.4, S.152-162
    Type
    a
  4. Lustig, G.: ¬Das Projekt WAI : Wörterbuchentwicklung für automatisches Indexing (1982) 0.02
    0.01769935 = product of:
      0.0353987 = sum of:
        0.0353987 = product of:
          0.053098045 = sum of:
            0.009410121 = weight(_text_:a in 33) [ClassicSimilarity], result of:
              0.009410121 = score(doc=33,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.17835285 = fieldWeight in 33, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.109375 = fieldNorm(doc=33)
            0.043687925 = weight(_text_:h in 33) [ClassicSimilarity], result of:
              0.043687925 = score(doc=33,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.38429362 = fieldWeight in 33, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.109375 = fieldNorm(doc=33)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Deutscher Dokumentartag 1981, Mainz, 5.-8.10.1981: Kleincomputer in Information und Dokumentation. Bearb.: H. Strohl-Goebel
    Type
    a
  5. Stock, W.G.: Textwortmethode (2000) 0.01
    0.012642393 = product of:
      0.025284786 = sum of:
        0.025284786 = product of:
          0.037927177 = sum of:
            0.0067215143 = weight(_text_:a in 3408) [ClassicSimilarity], result of:
              0.0067215143 = score(doc=3408,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.12739488 = fieldWeight in 3408, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3408)
            0.031205663 = weight(_text_:h in 3408) [ClassicSimilarity], result of:
              0.031205663 = score(doc=3408,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.27449545 = fieldWeight in 3408, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3408)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Auf dem Weg zur Informationskultur: Wa(h)re Information? Festschrift für Norbert Henrichs zum 65. Geburtstag, Hrsg.: T.A. Schröder
    Type
    a
  6. Stock, M.; Stock, W.G.: Literaturnachweis- und Terminologiedatenbank : die Erfassung von Fachliteratur und Fachterminologie eines Fachgebiets in einer kombinierten Datenbank (1991) 0.01
    0.012642393 = product of:
      0.025284786 = sum of:
        0.025284786 = product of:
          0.037927177 = sum of:
            0.0067215143 = weight(_text_:a in 3411) [ClassicSimilarity], result of:
              0.0067215143 = score(doc=3411,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.12739488 = fieldWeight in 3411, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3411)
            0.031205663 = weight(_text_:h in 3411) [ClassicSimilarity], result of:
              0.031205663 = score(doc=3411,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.27449545 = fieldWeight in 3411, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3411)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Nachrichten für Dokumentation. 42(1991) H.1, S.35-41
    Type
    a
  7. Gombocz, W.L.: Stichwort oder Schlagwort versus Textwort : Grazer und Düsseldorfer Philosophie-Dokumentation und -Information nach bzw. gemäß Norbert Henrichs (2000) 0.01
    0.012642393 = product of:
      0.025284786 = sum of:
        0.025284786 = product of:
          0.037927177 = sum of:
            0.0067215143 = weight(_text_:a in 3413) [ClassicSimilarity], result of:
              0.0067215143 = score(doc=3413,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.12739488 = fieldWeight in 3413, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3413)
            0.031205663 = weight(_text_:h in 3413) [ClassicSimilarity], result of:
              0.031205663 = score(doc=3413,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.27449545 = fieldWeight in 3413, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3413)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Auf dem Weg zur Informationskultur: Wa(h)re Information? Festschrift für Norbert Henrichs zum 65. Geburtstag, Hrsg.: T.A. Schröder
    Type
    a
  8. Volk, M.; Mittermaier, H.; Schurig, A.; Biedassek, T.: Halbautomatische Volltextanalyse, Datenbankaufbau und Document Retrieval (1992) 0.01
    0.012515331 = product of:
      0.025030661 = sum of:
        0.025030661 = product of:
          0.03754599 = sum of:
            0.0066539603 = weight(_text_:a in 2571) [ClassicSimilarity], result of:
              0.0066539603 = score(doc=2571,freq=4.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.12611452 = fieldWeight in 2571, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2571)
            0.03089203 = weight(_text_:h in 2571) [ClassicSimilarity], result of:
              0.03089203 = score(doc=2571,freq=4.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.27173662 = fieldWeight in 2571, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2571)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Datenanalyse, Klassifikation und Informationsverarbeitung: Methoden und Anwendungen in verschiedenen Fachgebieten. Hrsg.: H. Goebl u. M. Schader
    Type
    a
  9. Stock, W.G.: Textwortmethode : Norbert Henrichs zum 65. (3) (2000) 0.01
    0.010113914 = product of:
      0.020227827 = sum of:
        0.020227827 = product of:
          0.03034174 = sum of:
            0.0053772116 = weight(_text_:a in 4891) [ClassicSimilarity], result of:
              0.0053772116 = score(doc=4891,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.10191591 = fieldWeight in 4891, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4891)
            0.02496453 = weight(_text_:h in 4891) [ClassicSimilarity], result of:
              0.02496453 = score(doc=4891,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.21959636 = fieldWeight in 4891, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4891)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Password. 2000, H.7/8, S.26-35
    Type
    a
  10. Cheng, K.-H.: Automatic identification for topics of electronic documents (1997) 0.01
    0.009997789 = product of:
      0.019995578 = sum of:
        0.019995578 = product of:
          0.029993366 = sum of:
            0.008149404 = weight(_text_:a in 1811) [ClassicSimilarity], result of:
              0.008149404 = score(doc=1811,freq=6.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.1544581 = fieldWeight in 1811, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1811)
            0.021843962 = weight(_text_:h in 1811) [ClassicSimilarity], result of:
              0.021843962 = score(doc=1811,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.19214681 = fieldWeight in 1811, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1811)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Abstract
    With the rapid rise in numbers of electronic documents on the Internet, how to effectively assign topics to documents become an important issue. Current research in this area focuses on the behaviour of nouns in documents. Proposes, however, that nouns and verbs together contribute to the process of topic identification. Constructs a mathematical model taking into account the following factors: word importance, word frequency, word co-occurence, and word distance. Preliminary experiments ahow that the performance of the proposed model is equivalent to that of a human being
    Type
    a
  11. Rapke, K.: Automatische Indexierung von Volltexten für die Gruner+Jahr Pressedatenbank (2001) 0.01
    0.007585435 = product of:
      0.01517087 = sum of:
        0.01517087 = product of:
          0.022756305 = sum of:
            0.004032909 = weight(_text_:a in 6386) [ClassicSimilarity], result of:
              0.004032909 = score(doc=6386,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.07643694 = fieldWeight in 6386, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.046875 = fieldNorm(doc=6386)
            0.018723397 = weight(_text_:h in 6386) [ClassicSimilarity], result of:
              0.018723397 = score(doc=6386,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.16469726 = fieldWeight in 6386, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.046875 = fieldNorm(doc=6386)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    nfd Information - Wissenschaft und Praxis. 52(2001) H.5, S.251-262
    Type
    a
  12. SIGIR'92 : Proceedings of the 15th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (1992) 0.01
    0.0066777645 = product of:
      0.013355529 = sum of:
        0.013355529 = product of:
          0.020033292 = sum of:
            0.00911131 = weight(_text_:a in 6671) [ClassicSimilarity], result of:
              0.00911131 = score(doc=6671,freq=30.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.17268941 = fieldWeight in 6671, product of:
                  5.477226 = tf(freq=30.0), with freq of:
                    30.0 = termFreq=30.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=6671)
            0.010921981 = weight(_text_:h in 6671) [ClassicSimilarity], result of:
              0.010921981 = score(doc=6671,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.096073404 = fieldWeight in 6671, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=6671)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Content
    HARMAN, D.: Relevance feedback revisited; AALBERSBERG, I.J.: Incremental relevance feedback; TAGUE-SUTCLIFFE, J.: Measuring the informativeness of a retrieval process; LEWIS, D.D.: An evaluation of phrasal and clustered representations on a text categorization task; BLOSSEVILLE, M.J., G. HÉBRAIL, M.G. MONTEIL u. N. PÉNOT: Automatic document classification: natural language processing, statistical analysis, and expert system techniques used together; MASAND, B., G. LINOFF u. D. WALTZ: Classifying news stories using memory based reasoning; KEEN, E.M.: Term position ranking: some new test results; CROUCH, C.J. u. B. YANG: Experiments in automatic statistical thesaurus construction; GREFENSTETTE, G.: Use of syntactic context to produce term association lists for text retrieval; ANICK, P.G. u. R.A. FLYNN: Versioning of full-text information retrieval system; BURKOWSKI, F.J.: Retrieval activities in a database consisting of heterogeneous collections; DEERWESTER, S.C., K. WACLENA u. M. LaMAR: A textual object management system; NIE, J.-Y.:Towards a probabilistic modal logic for semantic-based information retrieval; WANG, A.W., S.K.M. WONG u. Y.Y. YAO: An analysis of vector space models based on computational geometry; BARTELL, B.T., G.W. COTTRELL u. R.K. BELEW: Latent semantic indexing is an optimal special case of multidimensional scaling; GLAVITSCH, U. u. P. SCHÄUBLE: A system for retrieving speech documents; MARGULIS, E.L.: N-Poisson document modelling; HESS, M.: An incrementally extensible document retrieval system based on linguistics and logical principles; COOPER, W.S., F.C. GEY u. D.P. DABNEY: Probabilistic retrieval based on staged logistic regression; FUHR, N.: Integration of probabilistic fact and text retrieval; CROFT, B., L.A. SMITH u. H. TURTLE: A loosely-coupled integration of a text retrieval system and an object-oriented database system; DUMAIS, S.T. u. J. NIELSEN: Automating the assignement of submitted manuscripts to reviewers; GOST, M.A. u. M. MASOTTI: Design of an OPAC database to permit different subject searching accesses; ROBERTSON, A.M. u. P. WILLETT: Searching for historical word forms in a database of 17th century English text using spelling correction methods; FAX, E.A., Q.F. CHEN u. L.S. HEATH: A faster algorithm for constructing minimal perfect hash functions; MOFFAT, A. u. J. ZOBEL: Parameterised compression for sparse bitmaps; GRANDI, F., P. TIBERIO u. P. Zezula: Frame-sliced patitioned parallel signature files; ALLEN, B.: Cognitive differences in end user searching of a CD-ROM index; SONNENWALD, D.H.: Developing a theory to guide the process of designing information retrieval systems; CUTTING, D.R., J.O. PEDERSEN, D. KARGER, u. J.W. TUKEY: Scatter/ Gather: a cluster-based approach to browsing large document collections; CHALMERS, M. u. P. CHITSON: Bead: Explorations in information visualization; WILLIAMSON, C. u. B. SHNEIDERMAN: The dynamic HomeFinder: evaluating dynamic queries in a real-estate information exploring system
  13. Larroche-Boutet, V.; Pöhl, K.: ¬Das Nominalsyntagna : über die Nutzbarmachung eines logico-semantischen Konzeptes für dokumentarische Fragestellungen (1993) 0.01
    0.0063211964 = product of:
      0.012642393 = sum of:
        0.012642393 = product of:
          0.018963588 = sum of:
            0.0033607571 = weight(_text_:a in 5282) [ClassicSimilarity], result of:
              0.0033607571 = score(doc=5282,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.06369744 = fieldWeight in 5282, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5282)
            0.015602832 = weight(_text_:h in 5282) [ClassicSimilarity], result of:
              0.015602832 = score(doc=5282,freq=2.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.13724773 = fieldWeight in 5282, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5282)
          0.6666667 = coord(2/3)
      0.5 = coord(1/2)
    
    Source
    Nachrichten für Dokumentation. 44(1993) H.5, S.269-276
    Type
    a
  14. Lorenz, S.: Konzeption und prototypische Realisierung einer begriffsbasierten Texterschließung (2006) 0.01
    0.0061995937 = product of:
      0.012399187 = sum of:
        0.012399187 = product of:
          0.03719756 = sum of:
            0.03719756 = weight(_text_:22 in 1746) [ClassicSimilarity], result of:
              0.03719756 = score(doc=1746,freq=2.0), product of:
                0.16023713 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045758117 = queryNorm
                0.23214069 = fieldWeight in 1746, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1746)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Date
    22. 3.2015 9:17:30
  15. Experimentelles und praktisches Information Retrieval : Festschrift für Gerhard Lustig (1992) 0.00
    0.004413147 = product of:
      0.008826294 = sum of:
        0.008826294 = product of:
          0.026478881 = sum of:
            0.026478881 = weight(_text_:h in 4) [ClassicSimilarity], result of:
              0.026478881 = score(doc=4,freq=4.0), product of:
                0.113683715 = queryWeight, product of:
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.045758117 = queryNorm
                0.2329171 = fieldWeight in 4, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  2.4844491 = idf(docFreq=10020, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Content
    Enthält die Beiträge: SALTON, G.: Effective text understanding in information retrieval; KRAUSE, J.: Intelligentes Information retrieval; FUHR, N.: Konzepte zur Gestaltung zukünftiger Information-Retrieval-Systeme; HÜTHER, H.: Überlegungen zu einem mathematischen Modell für die Type-Token-, die Grundform-Token und die Grundform-Type-Relation; KNORZ, G.: Automatische Generierung inferentieller Links in und zwischen Hyperdokumenten; KONRAD, E.: Zur Effektivitätsbewertung von Information-Retrieval-Systemen; HENRICHS, N.: Retrievalunterstützung durch automatisch generierte Wortfelder; LÜCK, W., W. RITTBERGER u. M. SCHWANTNER: Der Einsatz des Automatischen Indexierungs- und Retrieval-System (AIR) im Fachinformationszentrum Karlsruhe; REIMER, U.: Verfahren der Automatischen Indexierung. Benötigtes Vorwissen und Ansätze zu seiner automatischen Akquisition: Ein Überblick; ENDRES-NIGGEMEYER, B.: Dokumentrepräsentation: Ein individuelles prozedurales Modell des Abstracting, des Indexierens und Klassifizierens; SEELBACH, D.: Zur Entwicklung von zwei- und mehrsprachigen lexikalischen Datenbanken und Terminologiedatenbanken; ZIMMERMANN, H.: Der Einfluß der Sprachbarrieren in Europa und Möglichkeiten zu ihrer Minderung; LENDERS, W.: Wörter zwischen Welt und Wissen; PANYR, J.: Frames, Thesauri und automatische Klassifikation (Clusteranalyse): HAHN, U.: Forschungsstrategien und Erkenntnisinteressen in der anwendungsorientierten automatischen Sprachverarbeitung. Überlegungen zu einer ingenieurorientierten Computerlinguistik; KUHLEN, R.: Hypertext und Information Retrieval - mehr als Browsing und Suche.
  16. Driscoll, J.R.; Rajala, D.A.; Shaffer, W.H.: ¬The operation and performance of an artificially intelligent keywording system (1991) 0.00
    0.0022179869 = product of:
      0.0044359737 = sum of:
        0.0044359737 = product of:
          0.013307921 = sum of:
            0.013307921 = weight(_text_:a in 6681) [ClassicSimilarity], result of:
              0.013307921 = score(doc=6681,freq=16.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.25222903 = fieldWeight in 6681, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=6681)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Abstract
    Presents a new approach to text analysis for automating the key phrase indexing process, using artificial intelligence techniques. This mimics the behaviour of human experts by using a rule base consisting of insertion and deletion rules generated by subject-matter experts. The insertion rules are based on the idea that some phrases found in a text imply or trigger other phrases. The deletion rules apply to semantically ambiguous phrases where text presence alone does not determine appropriateness as a key phrase. The insertion and deletion rules are used to transform a list of found phrases to a list of key phrases for indexing a document. Statistical data are provided to demonstrate the performance of this expert rule based system
    Type
    a
  17. Pritchard-Schoch, T.: Natural language comes of age (1993) 0.00
    0.0017924039 = product of:
      0.0035848077 = sum of:
        0.0035848077 = product of:
          0.010754423 = sum of:
            0.010754423 = weight(_text_:a in 2570) [ClassicSimilarity], result of:
              0.010754423 = score(doc=2570,freq=8.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.20383182 = fieldWeight in 2570, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2570)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Abstract
    Discusses natural languages and the natural language implementations of Westlaw's full-text legal documents, Westlaw Is Natural. Natural language is not aritificial intelligence but a hybrid of linguistics, mathematics and statistics. Provides 3 classes of retrieval models. Explains how Westlaw processes an English query. Assesses WIN. Covers WIN enhancements; the natural language features of Congressional Quarterly's Washington Alert using a document for a query; the personal librarian front end search software and Dowquest from Dow Jones news/retrieval. Conmsiders whether natural language encourages fuzzy thinking and whether Boolean logic will still be needed
    Type
    a
  18. Salton, G.: Automatic processing of foreign language documents (1985) 0.00
    0.0016156512 = product of:
      0.0032313024 = sum of:
        0.0032313024 = product of:
          0.009693907 = sum of:
            0.009693907 = weight(_text_:a in 3650) [ClassicSimilarity], result of:
              0.009693907 = score(doc=3650,freq=26.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.18373153 = fieldWeight in 3650, product of:
                  5.0990195 = tf(freq=26.0), with freq of:
                    26.0 = termFreq=26.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3650)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Abstract
    The attempt to computerize a process, such as indexing, abstracting, classifying, or retrieving information, begins with an analysis of the process into its intellectual and nonintellectual components. That part of the process which is amenable to computerization is mechanical or algorithmic. What is not is intellectual or creative and requires human intervention. Gerard Salton has been an innovator, experimenter, and promoter in the area of mechanized information systems since the early 1960s. He has been particularly ingenious at analyzing the process of information retrieval into its algorithmic components. He received a doctorate in applied mathematics from Harvard University before moving to the computer science department at Cornell, where he developed a prototype automatic retrieval system called SMART. Working with this system he and his students contributed for over a decade to our theoretical understanding of the retrieval process. On a more practical level, they have contributed design criteria for operating retrieval systems. The following selection presents one of the early descriptions of the SMART system; it is valuable as it shows the direction automatic retrieval methods were to take beyond simple word-matching techniques. These include various word normalization techniques to improve recall, for instance, the separation of words into stems and affixes; the correlation and clustering, using statistical association measures, of related terms; and the identification, using a concept thesaurus, of synonymous, broader, narrower, and sibling terms. They include, as weIl, techniques, both linguistic and statistical, to deal with the thorny problem of how to automatically extract from texts index terms that consist of more than one word. They include weighting techniques and various documentrequest matching algorithms. Significant among the latter are those which produce a retrieval output of citations ranked in relevante order. During the 1970s, Salton and his students went an to further refine these various techniques, particularly the weighting and statistical association measures. Many of their early innovations seem commonplace today. Some of their later techniques are still ahead of their time and await technological developments for implementation. The particular focus of the selection that follows is an the evaluation of a particular component of the SMART system, a multilingual thesaurus. By mapping English language expressions and their German equivalents to a common concept number, the thesaurus permitted the automatic processing of German language documents against English language queries and vice versa. The results of the evaluation, as it turned out, were somewhat inconclusive. However, this SMART experiment suggested in a bold and optimistic way how one might proceed to answer such complex questions as What is meant by retrieval language compatability? How it is to be achieved, and how evaluated?
    Source
    Theory of subject analysis: a sourcebook. Ed.: L.M. Chan, et al
    Type
    a
  19. Kuhlen, R.: Morphologische Relationen durch Reduktionsalgorithmen (1974) 0.00
    0.0015683535 = product of:
      0.003136707 = sum of:
        0.003136707 = product of:
          0.009410121 = sum of:
            0.009410121 = weight(_text_:a in 4251) [ClassicSimilarity], result of:
              0.009410121 = score(doc=4251,freq=2.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.17835285 = fieldWeight in 4251, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4251)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Type
    a
  20. Malone, L.C.; Driscoll, J.R.; Pepe, J.W.: Modeling the performance of an automated keywording system (1991) 0.00
    0.0015522675 = product of:
      0.003104535 = sum of:
        0.003104535 = product of:
          0.009313605 = sum of:
            0.009313605 = weight(_text_:a in 6682) [ClassicSimilarity], result of:
              0.009313605 = score(doc=6682,freq=6.0), product of:
                0.052761257 = queryWeight, product of:
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.045758117 = queryNorm
                0.17652355 = fieldWeight in 6682, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.153047 = idf(docFreq=37942, maxDocs=44218)
                  0.0625 = fieldNorm(doc=6682)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Abstract
    Presents a model for predicting the performance of a computerised keyword assigning and indexing system. Statistical procedures were investigated in order to protect against incorrect keywording by the system behaving as an expert system designed to mimic the behaviour of human keyword indexers and representing lessons learned from military exercises and operations
    Type
    a