Search (24 results, page 1 of 2)

  • × theme_ss:"Sprachretrieval"
  1. Lange, H.R.: Speech synthesis and speech recognition : tomorrow's human-computer interface? (1993) 0.02
    0.018493315 = product of:
      0.03698663 = sum of:
        0.018330941 = weight(_text_:information in 7224) [ClassicSimilarity], result of:
          0.018330941 = score(doc=7224,freq=4.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.21943474 = fieldWeight in 7224, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=7224)
        0.01865569 = product of:
          0.03731138 = sum of:
            0.03731138 = weight(_text_:technology in 7224) [ClassicSimilarity], result of:
              0.03731138 = score(doc=7224,freq=2.0), product of:
                0.1417311 = queryWeight, product of:
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.047586527 = queryNorm
                0.2632547 = fieldWeight in 7224, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.0625 = fieldNorm(doc=7224)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Imprint
    Medford, NJ : Learned Information
    Source
    Annual review of information science and technology. 28(1993), S.153-185
  2. Srihari, R.K.: Using speech input for image interpretation, annotation, and retrieval (1997) 0.02
    0.016545078 = product of:
      0.033090156 = sum of:
        0.013748205 = weight(_text_:information in 764) [ClassicSimilarity], result of:
          0.013748205 = score(doc=764,freq=4.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.16457605 = fieldWeight in 764, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=764)
        0.019341951 = product of:
          0.038683902 = sum of:
            0.038683902 = weight(_text_:22 in 764) [ClassicSimilarity], result of:
              0.038683902 = score(doc=764,freq=2.0), product of:
                0.16663991 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.047586527 = queryNorm
                0.23214069 = fieldWeight in 764, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=764)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Explores the interaction of textual and photographic information in an integrated text and image database environment and describes 3 different applications involving the exploitation of linguistic context in vision. Describes the practical application of these ideas in working systems. PICTION uses captions to identify human faces in a photograph, wile Show&Tell is a multimedia system for semi automatic image annotation. The system combines advances in speech recognition, natural language processing and image understanding to assist in image annotation and enhance image retrieval capabilities. Presents an extension of this work to video annotation and retrieval
    Date
    22. 9.1997 19:16:05
    Imprint
    Urbana-Champaign, IL : Illinois University at Urbana-Champaign, Department of Library and Information Science
  3. Lin, J.; Katz, B.: Building a reusable test collection for question answering (2006) 0.02
    0.015414905 = product of:
      0.03082981 = sum of:
        0.016838044 = weight(_text_:information in 5045) [ClassicSimilarity], result of:
          0.016838044 = score(doc=5045,freq=6.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.20156369 = fieldWeight in 5045, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=5045)
        0.013991767 = product of:
          0.027983533 = sum of:
            0.027983533 = weight(_text_:technology in 5045) [ClassicSimilarity], result of:
              0.027983533 = score(doc=5045,freq=2.0), product of:
                0.1417311 = queryWeight, product of:
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.047586527 = queryNorm
                0.19744103 = fieldWeight in 5045, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5045)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    In contrast to traditional information retrieval systems, which return ranked lists of documents that users must manually browse through, a question answering system attempts to directly answer natural language questions posed by the user. Although such systems possess language-processing capabilities, they still rely on traditional document retrieval techniques to generate an initial candidate set of documents. In this article, the authors argue that document retrieval for question answering represents a task different from retrieving documents in response to more general retrospective information needs. Thus, to guide future system development, specialized question answering test collections must be constructed. They show that the current evaluation resources have major shortcomings; to remedy the situation, they have manually created a small, reusable question answering test collection for research purposes. In this article they describe their methodology for building this test collection and discuss issues they encountered regarding the notion of "answer correctness."
    Source
    Journal of the American Society for Information Science and Technology. 57(2006) no.7, S.851-861
  4. Pomerantz, J.: ¬A linguistic analysis of question taxonomies (2005) 0.01
    0.01383271 = product of:
      0.02766542 = sum of:
        0.011341691 = weight(_text_:information in 3465) [ClassicSimilarity], result of:
          0.011341691 = score(doc=3465,freq=2.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.13576832 = fieldWeight in 3465, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3465)
        0.016323728 = product of:
          0.032647457 = sum of:
            0.032647457 = weight(_text_:technology in 3465) [ClassicSimilarity], result of:
              0.032647457 = score(doc=3465,freq=2.0), product of:
                0.1417311 = queryWeight, product of:
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.047586527 = queryNorm
                0.23034787 = fieldWeight in 3465, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3465)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Source
    Journal of the American Society for Information Science and Technology. 56(2005) no.7, S.715-728
  5. Radev, D.; Fan, W.; Qu, H.; Wu, H.; Grewal, A.: Probabilistic question answering on the Web (2005) 0.01
    0.011856608 = product of:
      0.023713216 = sum of:
        0.00972145 = weight(_text_:information in 3455) [ClassicSimilarity], result of:
          0.00972145 = score(doc=3455,freq=2.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.116372846 = fieldWeight in 3455, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3455)
        0.013991767 = product of:
          0.027983533 = sum of:
            0.027983533 = weight(_text_:technology in 3455) [ClassicSimilarity], result of:
              0.027983533 = score(doc=3455,freq=2.0), product of:
                0.1417311 = queryWeight, product of:
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.047586527 = queryNorm
                0.19744103 = fieldWeight in 3455, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3455)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Source
    Journal of the American Society for Information Science and Technology. 56(2005) no.6, S.571-583
  6. Kruschwitz, U.; AI-Bakour, H.: Users want more sophisticated search assistants : results of a task-based evaluation (2005) 0.01
    0.0098805055 = product of:
      0.019761011 = sum of:
        0.008101207 = weight(_text_:information in 4575) [ClassicSimilarity], result of:
          0.008101207 = score(doc=4575,freq=2.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.09697737 = fieldWeight in 4575, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4575)
        0.011659805 = product of:
          0.02331961 = sum of:
            0.02331961 = weight(_text_:technology in 4575) [ClassicSimilarity], result of:
              0.02331961 = score(doc=4575,freq=2.0), product of:
                0.1417311 = queryWeight, product of:
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.047586527 = queryNorm
                0.16453418 = fieldWeight in 4575, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4575)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Source
    Journal of the American Society for Information Science and Technology. 56(2005) no.13, S.1377-1393
  7. Hannabuss, S.: Dialogue and the search for information (1989) 0.01
    0.0072459406 = product of:
      0.028983762 = sum of:
        0.028983762 = weight(_text_:information in 2590) [ClassicSimilarity], result of:
          0.028983762 = score(doc=2590,freq=10.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.3469568 = fieldWeight in 2590, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=2590)
      0.25 = coord(1/4)
    
    Abstract
    Knowledge of conversation theory and speech act assists us to understand how people search for information. Dialogue embodies meanings and intentionalities, and represents epistemic inquiry. There are implications for the information-processing model of cognitive psychology. Question formulation (erotetics) and turn-taking play important roles in eliciting information, while discourse analysis furnishes us with information about people's categorising, recall, and semantic skills
  8. Keller, F.: How do humans deal with ungrammatical input? : Experimental evidence and computational modelling (1996) 0.01
    0.0069958833 = product of:
      0.027983533 = sum of:
        0.027983533 = product of:
          0.055967066 = sum of:
            0.055967066 = weight(_text_:technology in 7293) [ClassicSimilarity], result of:
              0.055967066 = score(doc=7293,freq=2.0), product of:
                0.1417311 = queryWeight, product of:
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.047586527 = queryNorm
                0.39488205 = fieldWeight in 7293, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.09375 = fieldNorm(doc=7293)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Source
    Natural language processing and speech technology: Results of the 3rd KONVENS Conference, Bielefeld, October 1996. Ed.: D. Gibbon
  9. Marx, J.: ¬Die '¬Computer-Talk-These' in der Sprachgenerierung : Hinweise zur Gestaltung natürlichsprachlicher Zustandsanzeigen in multimodalen Informationssystemen (1996) 0.01
    0.0069958833 = product of:
      0.027983533 = sum of:
        0.027983533 = product of:
          0.055967066 = sum of:
            0.055967066 = weight(_text_:technology in 7294) [ClassicSimilarity], result of:
              0.055967066 = score(doc=7294,freq=2.0), product of:
                0.1417311 = queryWeight, product of:
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.047586527 = queryNorm
                0.39488205 = fieldWeight in 7294, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.09375 = fieldNorm(doc=7294)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Source
    Natural language processing and speech technology: Results of the 3rd KONVENS Conference, Bielefeld, October 1996. Ed.: D. Gibbon
  10. Schultz, T.; Soltau, H.: Automatische Identifizierung spontan gesprochener Sprachen mit neuronalen Netzen (1996) 0.01
    0.0069958833 = product of:
      0.027983533 = sum of:
        0.027983533 = product of:
          0.055967066 = sum of:
            0.055967066 = weight(_text_:technology in 7295) [ClassicSimilarity], result of:
              0.055967066 = score(doc=7295,freq=2.0), product of:
                0.1417311 = queryWeight, product of:
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.047586527 = queryNorm
                0.39488205 = fieldWeight in 7295, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  2.978387 = idf(docFreq=6114, maxDocs=44218)
                  0.09375 = fieldNorm(doc=7295)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Source
    Natural language processing and speech technology: Results of the 3rd KONVENS Conference, Bielefeld, October 1996. Ed.: D. Gibbon
  11. Peters, B.F.: Online searching using speech as a man / machine interface (1989) 0.01
    0.006480966 = product of:
      0.025923865 = sum of:
        0.025923865 = weight(_text_:information in 4637) [ClassicSimilarity], result of:
          0.025923865 = score(doc=4637,freq=2.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.3103276 = fieldWeight in 4637, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.125 = fieldNorm(doc=4637)
      0.25 = coord(1/4)
    
    Source
    Information processing and management. 25(1989), S.391-406
  12. Wittbrock, M.J.; Hauptmann, A.G.: Speech recognition for a digital video library (1998) 0.00
    0.0049609565 = product of:
      0.019843826 = sum of:
        0.019843826 = weight(_text_:information in 873) [ClassicSimilarity], result of:
          0.019843826 = score(doc=873,freq=12.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.23754507 = fieldWeight in 873, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=873)
      0.25 = coord(1/4)
    
    Abstract
    The standard method for making the full content of audio and video material searchable is to annotate it with human-generated meta-data that describes the content in a way that search can understand, as is done in the creation of multimedia CD-ROMs. However, for the huge amounts of data that could usefully be included in digital video and audio libraries, the cost of producing the meta-data is prohibitive. In the Informedia Digital Video Library, the production of the meta-data supporting the library interface is automated using techniques derived from artificial intelligence (AI) research. By applying speech recognition together with natural language processing, information retrieval, and image analysis, an interface has been prduced that helps users locate the information they want, and navigate or browse the digital video library more effectively. Specific interface components include automatc titles, filmstrips, video skims, word location marking, and representative frames for shots. Both the user interface and the information retrieval engine within Informedia are designed for use with automatically derived meta-data, much of which depends on speech recognition for its production. Some experimental information retrieval results will be given, supporting a basic premise of the Informedia project: That speech recognition generated transcripts can make multimedia material searchable. The Informedia project emphasizes the integration of speech recognition, image processing, natural language processing, and information retrieval to compensate for deficiencies in these individual technologies
    Source
    Journal of the American Society for Information Science. 49(1998) no.7, S.619-632
  13. Voorhees, E.M.: Question answering in TREC (2005) 0.00
    0.004860725 = product of:
      0.0194429 = sum of:
        0.0194429 = weight(_text_:information in 6487) [ClassicSimilarity], result of:
          0.0194429 = score(doc=6487,freq=2.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.23274569 = fieldWeight in 6487, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.09375 = fieldNorm(doc=6487)
      0.25 = coord(1/4)
    
    Source
    TREC: experiment and evaluation in information retrieval. Ed.: E.M. Voorhees, u. D.K. Harman
  14. Tartakovski, O.; Shramko, M.: Implementierung eines Werkzeugs zur Sprachidentifikation in mono- und multilingualen Texten (2006) 0.00
    0.004009893 = product of:
      0.016039573 = sum of:
        0.016039573 = weight(_text_:information in 5978) [ClassicSimilarity], result of:
          0.016039573 = score(doc=5978,freq=4.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.1920054 = fieldWeight in 5978, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5978)
      0.25 = coord(1/4)
    
    Abstract
    Die Identifikation der Sprache bzw. der Sprachen in Textdokumenten ist einer der wichtigsten Schritte maschineller Textverarbeitung für das Information Retrieval. Der vorliegende Artikel stellt Langldent vor, ein System zur Sprachidentifikation von mono- und multilingualen elektronischen Textdokumenten. Das System bietet sowohl eine Auswahl von gängigen Algorithmen für die Sprachidentifikation monolingualer Textdokumente als auch einen neuen Algorithmus für die Sprachidentifikation multilingualer Textdokumente.
    Source
    Effektive Information Retrieval Verfahren in Theorie und Praxis: ausgewählte und erweiterte Beiträge des Vierten Hildesheimer Evaluierungs- und Retrievalworkshop (HIER 2005), Hildesheim, 20.7.2005. Hrsg.: T. Mandl u. C. Womser-Hacker
  15. Young, C.W.; Eastman, C.M.; Oakman, R.L.: ¬An analysis of ill-formed input in natural language queries to document retrieval systems (1991) 0.00
    0.0034370513 = product of:
      0.013748205 = sum of:
        0.013748205 = weight(_text_:information in 5263) [ClassicSimilarity], result of:
          0.013748205 = score(doc=5263,freq=4.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.16457605 = fieldWeight in 5263, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=5263)
      0.25 = coord(1/4)
    
    Abstract
    Natrual language document retrieval queries from the Thomas Cooper Library, South Carolina Univ. were analysed in oder to investigate the frequency of various types of ill-formed input, such as spelling errors, cooccurrence violations, conjunctions, ellipsis, and missing or incorrect punctuation. Users were requested to write out their requests for information in complete sentences on the form normally used by the library. The primary reason for analysing ill-formed inputs was to determine whether there is a significant need to study ill-formed inputs in detail. Results indicated that most of the queries were sentence fragments and that many of them contained some type of ill-formed input. Conjunctions caused the most problems. The next most serious problem was caused by punctuation errors. Spelling errors occured in a small number of queries. The remaining types of ill-formed input considered, allipsis and cooccurrence violations, were not found in the queries
    Source
    Information processing and management. 27(1991) no.6, S.615-622
  16. Thompson, L.A.; Ogden, W.C.: Visible speech improves human language understanding : implications for speech processing systems (1995) 0.00
    0.003240483 = product of:
      0.012961932 = sum of:
        0.012961932 = weight(_text_:information in 3883) [ClassicSimilarity], result of:
          0.012961932 = score(doc=3883,freq=2.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.1551638 = fieldWeight in 3883, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=3883)
      0.25 = coord(1/4)
    
    Theme
    Information
  17. Burke, R.D.: Question answering from frequently asked question files : experiences with the FAQ Finder System (1997) 0.00
    0.003240483 = product of:
      0.012961932 = sum of:
        0.012961932 = weight(_text_:information in 1191) [ClassicSimilarity], result of:
          0.012961932 = score(doc=1191,freq=2.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.1551638 = fieldWeight in 1191, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=1191)
      0.25 = coord(1/4)
    
    Abstract
    Describes FAQ Finder, a natural language question-answering system that uses files of frequently asked questions as its knowledge base. Unlike information retrieval approaches that rely on a purely lexical metric of similarity between query and document, FAQ Finder uses a semantic knowledge base (Wordnet) to improve its ability to match question and answer. Includes results from an evaluation of the system's performance and shows that a combination of semantic and statistical techniques works better than any single approach
  18. Nhongkai, S.N.; Bentz, H.-J.: Bilinguale Suche mittels Konzeptnetzen (2006) 0.00
    0.003240483 = product of:
      0.012961932 = sum of:
        0.012961932 = weight(_text_:information in 3914) [ClassicSimilarity], result of:
          0.012961932 = score(doc=3914,freq=2.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.1551638 = fieldWeight in 3914, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0625 = fieldNorm(doc=3914)
      0.25 = coord(1/4)
    
    Source
    Effektive Information Retrieval Verfahren in Theorie und Praxis: ausgewählte und erweiterte Beiträge des Vierten Hildesheimer Evaluierungs- und Retrievalworkshop (HIER 2005), Hildesheim, 20.7.2005. Hrsg.: T. Mandl u. C. Womser-Hacker
  19. Sparck Jones, K.; Jones, G.J.F.; Foote, J.T.; Young, S.J.: Experiments in spoken document retrieval (1996) 0.00
    0.0028354228 = product of:
      0.011341691 = sum of:
        0.011341691 = weight(_text_:information in 1951) [ClassicSimilarity], result of:
          0.011341691 = score(doc=1951,freq=2.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.13576832 = fieldWeight in 1951, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1951)
      0.25 = coord(1/4)
    
    Source
    Information processing and management. 32(1996) no.4, S.399-417
  20. Rösener, C.: ¬Die Stecknadel im Heuhaufen : Natürlichsprachlicher Zugang zu Volltextdatenbanken (2005) 0.00
    0.0028063408 = product of:
      0.011225363 = sum of:
        0.011225363 = weight(_text_:information in 548) [ClassicSimilarity], result of:
          0.011225363 = score(doc=548,freq=6.0), product of:
            0.083537094 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.047586527 = queryNorm
            0.1343758 = fieldWeight in 548, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=548)
      0.25 = coord(1/4)
    
    Abstract
    Die Möglichkeiten, die der heutigen Informations- und Wissensgesellschaft für die Beschaffung und den Austausch von Information zur Verfügung stehen, haben kurioserweise gleichzeitig ein immer akuter werdendes, neues Problem geschaffen: Es wird für jeden Einzelnen immer schwieriger, aus der gewaltigen Fülle der angebotenen Informationen die tatsächlich relevanten zu selektieren. Diese Arbeit untersucht die Möglichkeit, mit Hilfe von natürlichsprachlichen Schnittstellen den Zugang des Informationssuchenden zu Volltextdatenbanken zu verbessern. Dabei werden zunächst die wissenschaftlichen Fragestellungen ausführlich behandelt. Anschließend beschreibt der Autor verschiedene Lösungsansätze und stellt anhand einer natürlichsprachlichen Schnittstelle für den Brockhaus Multimedial 2004 deren erfolgreiche Implementierung vor
    Content
    Enthält die Kapitel: 2: Wissensrepräsentation 2.1 Deklarative Wissensrepräsentation 2.2 Klassifikationen des BMM 2.3 Thesauri und Ontologien: existierende kommerzielle Software 2.4 Erstellung eines Thesaurus im Rahmen des LeWi-Projektes 3: Analysekomponenten 3.1 Sprachliche Phänomene in der maschinellen Textanalyse 3.2 Analysekomponenten: Lösungen und Forschungsansätze 3.3 Die Analysekomponenten im LeWi-Projekt 4: Information Retrieval 4.1 Grundlagen des Information Retrieval 4.2 Automatische Indexierungsmethoden und -verfahren 4.3 Automatische Indexierung des BMM im Rahmen des LeWi-Projektes 4.4 Suchstrategien und Suchablauf im LeWi-Kontext