Search (237 results, page 1 of 12)

  • × theme_ss:"Wissensrepräsentation"
  1. Xiong, C.: Knowledge based text representations for information retrieval (2016) 0.14
    0.13877001 = product of:
      0.27754003 = sum of:
        0.04785113 = product of:
          0.14355339 = sum of:
            0.14355339 = weight(_text_:3a in 5820) [ClassicSimilarity], result of:
              0.14355339 = score(doc=5820,freq=2.0), product of:
                0.38313732 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.045191888 = queryNorm
                0.3746787 = fieldWeight in 5820, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5820)
          0.33333334 = coord(1/3)
        0.20301515 = weight(_text_:2f in 5820) [ClassicSimilarity], result of:
          0.20301515 = score(doc=5820,freq=4.0), product of:
            0.38313732 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.045191888 = queryNorm
            0.5298757 = fieldWeight in 5820, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
        0.026673725 = weight(_text_:computer in 5820) [ClassicSimilarity], result of:
          0.026673725 = score(doc=5820,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.16150802 = fieldWeight in 5820, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.03125 = fieldNorm(doc=5820)
      0.5 = coord(3/6)
    
    Content
    Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Language and Information Technologies. Vgl.: https%3A%2F%2Fwww.cs.cmu.edu%2F~cx%2Fpapers%2Fknowledge_based_text_representation.pdf&usg=AOvVaw0SaTSvhWLTh__Uz_HtOtl3.
    Imprint
    Pittsburgh, PA : Carnegie Mellon University, School of Computer Science, Language Technologies Institute
  2. Zeng, Q.; Yu, M.; Yu, W.; Xiong, J.; Shi, Y.; Jiang, M.: Faceted hierarchy : a new graph type to organize scientific concepts and a construction method (2019) 0.10
    0.09570226 = product of:
      0.28710678 = sum of:
        0.071776696 = product of:
          0.21533008 = sum of:
            0.21533008 = weight(_text_:3a in 400) [ClassicSimilarity], result of:
              0.21533008 = score(doc=400,freq=2.0), product of:
                0.38313732 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.045191888 = queryNorm
                0.56201804 = fieldWeight in 400, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=400)
          0.33333334 = coord(1/3)
        0.21533008 = weight(_text_:2f in 400) [ClassicSimilarity], result of:
          0.21533008 = score(doc=400,freq=2.0), product of:
            0.38313732 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.045191888 = queryNorm
            0.56201804 = fieldWeight in 400, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=400)
      0.33333334 = coord(2/6)
    
    Content
    Vgl.: https%3A%2F%2Faclanthology.org%2FD19-5317.pdf&usg=AOvVaw0ZZFyq5wWTtNTvNkrvjlGA.
  3. Stojanovic, N.: Ontology-based Information Retrieval : methods and tools for cooperative query answering (2005) 0.06
    0.06380151 = product of:
      0.19140452 = sum of:
        0.04785113 = product of:
          0.14355339 = sum of:
            0.14355339 = weight(_text_:3a in 701) [ClassicSimilarity], result of:
              0.14355339 = score(doc=701,freq=2.0), product of:
                0.38313732 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.045191888 = queryNorm
                0.3746787 = fieldWeight in 701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03125 = fieldNorm(doc=701)
          0.33333334 = coord(1/3)
        0.14355339 = weight(_text_:2f in 701) [ClassicSimilarity], result of:
          0.14355339 = score(doc=701,freq=2.0), product of:
            0.38313732 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.045191888 = queryNorm
            0.3746787 = fieldWeight in 701, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03125 = fieldNorm(doc=701)
      0.33333334 = coord(2/6)
    
    Content
    Vgl.: http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F1627&ei=tAtYUYrBNoHKtQb3l4GYBw&usg=AFQjCNHeaxKkKU3-u54LWxMNYGXaaDLCGw&sig2=8WykXWQoDKjDSdGtAakH2Q&bvm=bv.44442042,d.Yms.
  4. Innovations and advanced techniques in systems, computing sciences and software engineering (2008) 0.06
    0.058486085 = product of:
      0.17545825 = sum of:
        0.105437145 = weight(_text_:computer in 4319) [ClassicSimilarity], result of:
          0.105437145 = score(doc=4319,freq=20.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.63841647 = fieldWeight in 4319, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4319)
        0.0700211 = weight(_text_:network in 4319) [ClassicSimilarity], result of:
          0.0700211 = score(doc=4319,freq=4.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.34791988 = fieldWeight in 4319, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4319)
      0.33333334 = coord(2/6)
    
    Abstract
    Innovations and Advanced Techniques in Systems, Computing Sciences and Software Engineering includes a set of rigorously reviewed world-class manuscripts addressing and detailing state-of-the-art research projects in the areas of Computer Science, Software Engineering, Computer Engineering, and Systems Engineering and Sciences. Innovations and Advanced Techniques in Systems, Computing Sciences and Software Engineering includes selected papers form the conference proceedings of the International Conference on Systems, Computing Sciences and Software Engineering (SCSS 2007) which was part of the International Joint Conferences on Computer, Information and Systems Sciences and Engineering (CISSE 2007).
    Content
    Inhalt: Image and Pattern Recognition: Compression, Image processing, Signal Processing Architectures, Signal Processing for Communication, Signal Processing Implementation, Speech Compression, and Video Coding Architectures. Languages and Systems: Algorithms, Databases, Embedded Systems and Applications, File Systems and I/O, Geographical Information Systems, Kernel and OS Structures, Knowledge Based Systems, Modeling and Simulation, Object Based Software Engineering, Programming Languages, and Programming Models and tools. Parallel Processing: Distributed Scheduling, Multiprocessing, Real-time Systems, Simulation Modeling and Development, and Web Applications. New trends in computing: Computers for People of Special Needs, Fuzzy Inference, Human Computer Interaction, Incremental Learning, Internet-based Computing Models, Machine Intelligence, Natural Language Processing, Neural Networks, and Online Decision Support System
    LCSH
    Computer Science
    Computer Systems Organization and Communication Networks
    Computer network architectures
    Subject
    Computer Science
    Computer Systems Organization and Communication Networks
    Computer network architectures
  5. Giri, K.; Gokhale, P.: Developing a banking service ontology using Protégé, an open source software (2015) 0.05
    0.049743798 = product of:
      0.099487595 = sum of:
        0.033342157 = weight(_text_:computer in 2793) [ClassicSimilarity], result of:
          0.033342157 = score(doc=2793,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.20188503 = fieldWeight in 2793, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2793)
        0.049512394 = weight(_text_:network in 2793) [ClassicSimilarity], result of:
          0.049512394 = score(doc=2793,freq=2.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.2460165 = fieldWeight in 2793, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2793)
        0.016633041 = product of:
          0.033266082 = sum of:
            0.033266082 = weight(_text_:resources in 2793) [ClassicSimilarity], result of:
              0.033266082 = score(doc=2793,freq=2.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.20165458 = fieldWeight in 2793, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2793)
          0.5 = coord(1/2)
      0.5 = coord(3/6)
    
    Abstract
    Computers have transformed from single isolated devices to entry points into a worldwide network of information exchange. Consequently, support in the exchange of data, information, and knowledge is becoming the key issue in computer technology today. The increasing volume of data available on the Web makes information retrieval a tedious and difficult task. Researchers are now exploring the possibility of creating a semantic web, in which meaning is made explicit, allowing machines to process and integrate web resources intelligently. The vision of the semantic web introduces the next generation of the Web by establishing a layer of machine-understandable data. The success of the semantic web depends on the easy creation, integration and use of semantic data, which will depend on web ontology. The faceted approach towards analyzing and representing knowledge given by S R Ranganathan would be useful in this regard. Ontology development in different fields is one such area where this approach given by Ranganathan could be applied. This paper presents a case of developing ontology for the field of banking.
  6. Panzer, M.: Towards the "webification" of controlled subject vocabulary : a case study involving the Dewey Decimal Classification (2007) 0.05
    0.045314007 = product of:
      0.13594201 = sum of:
        0.066624664 = weight(_text_:services in 538) [ClassicSimilarity], result of:
          0.066624664 = score(doc=538,freq=4.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.40155616 = fieldWeight in 538, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.0546875 = fieldNorm(doc=538)
        0.06931735 = weight(_text_:network in 538) [ClassicSimilarity], result of:
          0.06931735 = score(doc=538,freq=2.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.3444231 = fieldWeight in 538, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.0546875 = fieldNorm(doc=538)
      0.33333334 = coord(2/6)
    
    Abstract
    The presentation will briefly introduce a series of major principles for bringing subject terminology to the network level. A closer look at one KOS in particular, the Dewey Decimal Classification, should help to gain more insight into the perceived difficulties and potential benefits of building taxonomy services out and on top of classic large-scale vocabularies or taxonomies.
    Content
    Präsentation während der Veranstaltung "Networked Knowledge Organization Systems and Services: The 6th European Networked Knowledge Organization Systems (NKOS) Workshop, Workshop at the 11th ECDL Conference, Budapest, Hungary, September 21st 2007".
  7. Gendt, M. van; Isaac, I.; Meij, L. van der; Schlobach, S.: Semantic Web techniques for multiple views on heterogeneous collections : a case study (2006) 0.04
    0.03888905 = product of:
      0.11666715 = sum of:
        0.040010586 = weight(_text_:computer in 2418) [ClassicSimilarity], result of:
          0.040010586 = score(doc=2418,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.24226204 = fieldWeight in 2418, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.046875 = fieldNorm(doc=2418)
        0.076656565 = sum of:
          0.0399193 = weight(_text_:resources in 2418) [ClassicSimilarity], result of:
            0.0399193 = score(doc=2418,freq=2.0), product of:
              0.16496566 = queryWeight, product of:
                3.650338 = idf(docFreq=3122, maxDocs=44218)
                0.045191888 = queryNorm
              0.2419855 = fieldWeight in 2418, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.650338 = idf(docFreq=3122, maxDocs=44218)
                0.046875 = fieldNorm(doc=2418)
          0.036737263 = weight(_text_:22 in 2418) [ClassicSimilarity], result of:
            0.036737263 = score(doc=2418,freq=2.0), product of:
              0.1582543 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.045191888 = queryNorm
              0.23214069 = fieldWeight in 2418, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2418)
      0.33333334 = coord(2/6)
    
    Abstract
    Integrated digital access to multiple collections is a prominent issue for many Cultural Heritage institutions. The metadata describing diverse collections must be interoperable, which requires aligning the controlled vocabularies that are used to annotate objects from these collections. In this paper, we present an experiment where we match the vocabularies of two collections by applying the Knowledge Representation techniques established in recent Semantic Web research. We discuss the steps that are required for such matching, namely formalising the initial resources using Semantic Web languages, and running ontology mapping tools on the resulting representations. In addition, we present a prototype that enables the user to browse the two collections using the obtained alignment while still providing her with the original vocabulary structures.
    Series
    Lecture notes in computer science; vol.4172
    Source
    Research and advanced technology for digital libraries : 10th European conference, proceedings / ECDL 2006, Alicante, Spain, September 17 - 22, 2006
  8. ¬The Semantic Web : research and applications ; second European Semantic WebConference, ESWC 2005, Heraklion, Crete, Greece, May 29 - June 1, 2005 ; proceedings (2005) 0.04
    0.036560338 = product of:
      0.10968101 = sum of:
        0.040380646 = weight(_text_:services in 439) [ClassicSimilarity], result of:
          0.040380646 = score(doc=439,freq=2.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.2433798 = fieldWeight in 439, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.046875 = fieldNorm(doc=439)
        0.06930037 = weight(_text_:computer in 439) [ClassicSimilarity], result of:
          0.06930037 = score(doc=439,freq=6.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.41961014 = fieldWeight in 439, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.046875 = fieldNorm(doc=439)
      0.33333334 = coord(2/6)
    
    Abstract
    This book constitutes the refereed proceedings of the Second European Semantic Web Conference, ESWC 2005, heldin Heraklion, Crete, Greece in May/June 2005. The 48 revised full papers presented were carefully reviewed and selected from 148 submissions. The papers are organized in topical sections on semantic Web services, languages, ontologies, reasoning and querying, search and information retrieval, user and communities, natural language for the semantic Web, annotation tools, and semantic Web applications.
    LCSH
    Computer Communication Networks
    Series
    Lecture Notes in Computer Science ; 3532
    Subject
    Computer Communication Networks
  9. Castellanos Ardila, J.P.: Investigation of an OSLC-domain targeting ISO 26262 : focus on the left side of the software V-model (2016) 0.03
    0.033450294 = product of:
      0.06690059 = sum of:
        0.02692043 = weight(_text_:services in 5819) [ClassicSimilarity], result of:
          0.02692043 = score(doc=5819,freq=2.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.1622532 = fieldWeight in 5819, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.03125 = fieldNorm(doc=5819)
        0.026673725 = weight(_text_:computer in 5819) [ClassicSimilarity], result of:
          0.026673725 = score(doc=5819,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.16150802 = fieldWeight in 5819, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.03125 = fieldNorm(doc=5819)
        0.013306432 = product of:
          0.026612865 = sum of:
            0.026612865 = weight(_text_:resources in 5819) [ClassicSimilarity], result of:
              0.026612865 = score(doc=5819,freq=2.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.16132367 = fieldWeight in 5819, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5819)
          0.5 = coord(1/2)
      0.5 = coord(3/6)
    
    Abstract
    Industries have adopted a standardized set of practices for developing their products. In the automotive domain, the provision of safety-compliant systems is guided by ISO 26262, a standard that specifies a set of requirements and recommendations for developing automotive safety-critical systems. For being in compliance with ISO 26262, the safety lifecycle proposed by the standard must be included in the development process of a vehicle. Besides, a safety case that shows that the system is acceptably safe has to be provided. The provision of a safety case implies the execution of a precise documentation process. This process makes sure that the work products are available and traceable. Further, the documentation management is defined in the standard as a mandatory activity and guidelines are proposed/imposed for its elaboration. It would be appropriate to point out that a well-documented safety lifecycle will provide the necessary inputs for the generation of an ISO 26262-compliant safety case. The OSLC (Open Services for Lifecycle Collaboration) standard and the maturing stack of semantic web technologies represent a promising integration platform for enabling semantic interoperability between the tools involved in the safety lifecycle. Tools for requirements, architecture, development management, among others, are expected to interact and shared data with the help of domains specifications created in OSLC. This thesis proposes the creation of an OSLC tool-chain infrastructure for sharing safety-related information, where fragments of safety information can be generated. The steps carried out during the elaboration of this master thesis consist in the identification, representation, and shaping of the RDF resources needed for the creation of a safety case. The focus of the thesis is limited to a tiny portion of the ISO 26262 left-hand side of the V-model, more exactly part 6 clause 8 of the standard: Software unit design and implementation. Regardless of the use of a restricted portion of the standard during the execution of this thesis, the findings can be extended to other parts, and the conclusions can be generalize. This master thesis is considered one of the first steps towards the provision of an OSLC-based and ISO 26262-compliant methodological approach for representing and shaping the work products resulting from the execution of the safety lifecycle, documentation required in the conformation of an ISO-compliant safety case.
    Footnote
    Thesis for: Master of Computer Science with Specialization in Software Engineering. Advisor: Barbara Gallina.
  10. Drexel, G.: Knowledge engineering for intelligent information retrieval (2001) 0.03
    0.033141818 = product of:
      0.09942545 = sum of:
        0.040010586 = weight(_text_:computer in 4043) [ClassicSimilarity], result of:
          0.040010586 = score(doc=4043,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.24226204 = fieldWeight in 4043, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.046875 = fieldNorm(doc=4043)
        0.059414867 = weight(_text_:network in 4043) [ClassicSimilarity], result of:
          0.059414867 = score(doc=4043,freq=2.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.29521978 = fieldWeight in 4043, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.046875 = fieldNorm(doc=4043)
      0.33333334 = coord(2/6)
    
    Abstract
    This paper presents a clustered approach to designing an overall ontological model together with a general rule-based component that serves as a mapping device. By observational criteria, a multi-lingual team of experts excerpts concepts from general communication in the media. The team, then, finds equivalent expressions in English, German, French, and Spanish. On the basis of a set of ontological and lexical relations, a conceptual network is built up. Concepts are thought to be universal. Objects unique in time and space are identified by names and will be explained by the universals as their instances. Our approach relies on multi-relational descriptions of concepts. It provides a powerful tool for documentation and conceptual language learning. First and foremost, our multi-lingual, polyhierarchical ontology fills the gap of semantically-based information retrieval by generating enhanced and improved queries for internet search
    Series
    Lecture notes in computer science; vol.2004
  11. Quillian, M.R.: Word concepts : a theory and simulation of some basic semantic capabilities. (1967) 0.03
    0.033141818 = product of:
      0.09942545 = sum of:
        0.040010586 = weight(_text_:computer in 4414) [ClassicSimilarity], result of:
          0.040010586 = score(doc=4414,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.24226204 = fieldWeight in 4414, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.046875 = fieldNorm(doc=4414)
        0.059414867 = weight(_text_:network in 4414) [ClassicSimilarity], result of:
          0.059414867 = score(doc=4414,freq=2.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.29521978 = fieldWeight in 4414, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.046875 = fieldNorm(doc=4414)
      0.33333334 = coord(2/6)
    
    Abstract
    In order to discover design principles for a large memory that can enable it to serve as the base of knowledge underlying human-like language behavior, experiments with a model memory are being performed. This model is built up within a computer by "recoding" a body of information from an ordinary dictionary into a complex network of elements and associations interconnecting them. Then, the ability of a program to use the resulting model memory effectively for simulating human performance provides a test of its design. One simulation program, now running, is given the model memory and is required to compare and contrast the meanings of arbitrary pairs of English words. For each pair, the program locates any relevant semantic information within the model memory, draws inferences on the basis of this, and thereby discovers various relationships between the meanings of the two words. Finally, it creates English text to express its conclusions. The design principles embodied in the memory model, together with some of the methods used by the program, constitute a theory of how human memory for semantic and other conceptual material may be formatted, organized, and used.
  12. Frisch, A.M.; Allen, J.F.: Knowledge retrieval as limited inference (1982) 0.03
    0.033141818 = product of:
      0.09942545 = sum of:
        0.040010586 = weight(_text_:computer in 5804) [ClassicSimilarity], result of:
          0.040010586 = score(doc=5804,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.24226204 = fieldWeight in 5804, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.046875 = fieldNorm(doc=5804)
        0.059414867 = weight(_text_:network in 5804) [ClassicSimilarity], result of:
          0.059414867 = score(doc=5804,freq=2.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.29521978 = fieldWeight in 5804, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.046875 = fieldNorm(doc=5804)
      0.33333334 = coord(2/6)
    
    Abstract
    Artificial intelligence reasoning systems commonly employ a knowledge base module that stores a set of facts expressed in a representation language and provides facilities to retrieve these facts. A retriever could range from a simple pattern matcher to a complete logical inference system. In practice, most fall in between these extremes, providing some forms of inference but not others. Unfortunately, most of these retrievers are not precisely defined. We view knowledge retrieval as a limited form of inference operating on the stored facts. This paper is concerned with our method of using first-order predicate calculus to formally specify a limited inference mechanism and to a lesser extent with the techniques for producing an efficient program that meets the specification. Our ideas are illustrated by developing a simplified version of a retriever used in the knowledge base of the Rochester Dialog System. The interesting property of this retriever is that it perlorms typical semantic network inferences such as inheritance but not arbitrary logical inferences such as modus ponens.
    Series
    Lecture notes in computer science; vol 138
  13. Tudhope, D.; Hodge, G.: Terminology registries (2007) 0.03
    0.03263849 = product of:
      0.09791546 = sum of:
        0.06730108 = weight(_text_:services in 539) [ClassicSimilarity], result of:
          0.06730108 = score(doc=539,freq=2.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.405633 = fieldWeight in 539, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.078125 = fieldNorm(doc=539)
        0.030614385 = product of:
          0.06122877 = sum of:
            0.06122877 = weight(_text_:22 in 539) [ClassicSimilarity], result of:
              0.06122877 = score(doc=539,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.38690117 = fieldWeight in 539, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=539)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Content
    Präsentation während der Veranstaltung "Networked Knowledge Organization Systems and Services: The 6th European Networked Knowledge Organization Systems (NKOS) Workshop, Workshop at the 11th ECDL Conference, Budapest, Hungary, September 21st 2007".
    Date
    26.12.2011 13:22:07
  14. Nielsen, M.: Neuronale Netze : Alpha Go - Computer lernen Intuition (2018) 0.03
    0.0324329 = product of:
      0.0972987 = sum of:
        0.06668431 = weight(_text_:computer in 4523) [ClassicSimilarity], result of:
          0.06668431 = score(doc=4523,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.40377006 = fieldWeight in 4523, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.078125 = fieldNorm(doc=4523)
        0.030614385 = product of:
          0.06122877 = sum of:
            0.06122877 = weight(_text_:22 in 4523) [ClassicSimilarity], result of:
              0.06122877 = score(doc=4523,freq=2.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.38690117 = fieldWeight in 4523, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=4523)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Source
    Spektrum der Wissenschaft. 2018, H.1, S.22-27
  15. Baofu, P.: ¬The future of information architecture : conceiving a better way to understand taxonomy, network, and intelligence (2008) 0.03
    0.03118126 = product of:
      0.093543775 = sum of:
        0.0700211 = weight(_text_:network in 2257) [ClassicSimilarity], result of:
          0.0700211 = score(doc=2257,freq=4.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.34791988 = fieldWeight in 2257, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2257)
        0.023522673 = product of:
          0.047045346 = sum of:
            0.047045346 = weight(_text_:resources in 2257) [ClassicSimilarity], result of:
              0.047045346 = score(doc=2257,freq=4.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.28518265 = fieldWeight in 2257, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2257)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    The Future of Information Architecture examines issues surrounding why information is processed, stored and applied in the way that it has, since time immemorial. Contrary to the conventional wisdom held by many scholars in human history, the recurrent debate on the explanation of the most basic categories of information (eg space, time causation, quality, quantity) has been misconstrued, to the effect that there exists some deeper categories and principles behind these categories of information - with enormous implications for our understanding of reality in general. To understand this, the book is organised in to four main parts: Part I begins with the vital question concerning the role of information within the context of the larger theoretical debate in the literature. Part II provides a critical examination of the nature of data taxonomy from the main perspectives of culture, society, nature and the mind. Part III constructively invesitgates the world of information network from the main perspectives of culture, society, nature and the mind. Part IV proposes six main theses in the authors synthetic theory of information architecture, namely, (a) the first thesis on the simpleness-complicatedness principle, (b) the second thesis on the exactness-vagueness principle (c) the third thesis on the slowness-quickness principle (d) the fourth thesis on the order-chaos principle, (e) the fifth thesis on the symmetry-asymmetry principle, and (f) the sixth thesis on the post-human stage.
    LCSH
    Information resources
    Subject
    Information resources
  16. Cimiano, P.; Völker, J.; Studer, R.: Ontologies on demand? : a description of the state-of-the-art, applications, challenges and trends for ontology learning from text (2006) 0.03
    0.030559327 = product of:
      0.09167798 = sum of:
        0.057106856 = weight(_text_:services in 6014) [ClassicSimilarity], result of:
          0.057106856 = score(doc=6014,freq=4.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.344191 = fieldWeight in 6014, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.046875 = fieldNorm(doc=6014)
        0.034571126 = product of:
          0.06914225 = sum of:
            0.06914225 = weight(_text_:resources in 6014) [ClassicSimilarity], result of:
              0.06914225 = score(doc=6014,freq=6.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.4191312 = fieldWeight in 6014, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.046875 = fieldNorm(doc=6014)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Abstract
    Ontologies are nowadays used for many applications requiring data, services and resources in general to be interoperable and machine understandable. Such applications are for example web service discovery and composition, information integration across databases, intelligent search, etc. The general idea is that data and services are semantically described with respect to ontologies, which are formal specifications of a domain of interest, and can thus be shared and reused in a way such that the shared meaning specified by the ontology remains formally the same across different parties and applications. As the cost of creating ontologies is relatively high, different proposals have emerged for learning ontologies from structured and unstructured resources. In this article we examine the maturity of techniques for ontology learning from textual resources, addressing the question whether the state-of-the-art is mature enough to produce ontologies 'on demand'.
  17. ISO 25964 Thesauri and interoperability with other vocabularies (2008) 0.03
    0.029846275 = product of:
      0.05969255 = sum of:
        0.020005293 = weight(_text_:computer in 1169) [ClassicSimilarity], result of:
          0.020005293 = score(doc=1169,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.12113102 = fieldWeight in 1169, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1169)
        0.029707434 = weight(_text_:network in 1169) [ClassicSimilarity], result of:
          0.029707434 = score(doc=1169,freq=2.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.14760989 = fieldWeight in 1169, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1169)
        0.009979825 = product of:
          0.01995965 = sum of:
            0.01995965 = weight(_text_:resources in 1169) [ClassicSimilarity], result of:
              0.01995965 = score(doc=1169,freq=2.0), product of:
                0.16496566 = queryWeight, product of:
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.045191888 = queryNorm
                0.12099275 = fieldWeight in 1169, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.650338 = idf(docFreq=3122, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1169)
          0.5 = coord(1/2)
      0.5 = coord(3/6)
    
    Abstract
    T.1: Today's thesauri are mostly electronic tools, having moved on from the paper-based era when thesaurus standards were first developed. They are built and maintained with the support of software and need to integrate with other software, such as search engines and content management systems. Whereas in the past thesauri were designed for information professionals trained in indexing and searching, today there is a demand for vocabularies that untrained users will find to be intuitive. ISO 25964 makes the transition needed for the world of electronic information management. However, part 1 retains the assumption that human intellect is usually involved in the selection of indexing terms and in the selection of search terms. If both the indexer and the searcher are guided to choose the same term for the same concept, then relevant documents will be retrieved. This is the main principle underlying thesaurus design, even though a thesaurus built for human users may also be applied in situations where computers make the choices. Efficient exchange of data is a vital component of thesaurus management and exploitation. Hence the inclusion in this standard of recommendations for exchange formats and protocols. Adoption of these will facilitate interoperability between thesaurus management systems and the other computer applications, such as indexing and retrieval systems, that will utilize the data. Thesauri are typically used in post-coordinate retrieval systems, but may also be applied to hierarchical directories, pre-coordinate indexes and classification systems. Increasingly, thesaurus applications need to mesh with others, such as automatic categorization schemes, free-text search systems, etc. Part 2 of ISO 25964 describes additional types of structured vocabulary and gives recommendations to enable interoperation of the vocabularies at all stages of the information storage and retrieval process.
    T.2: The ability to identify and locate relevant information among vast collections and other resources is a major and pressing challenge today. Several different types of vocabulary are in use for this purpose. Some of the most widely used vocabularies were designed a hundred years ago and have been evolving steadily. A different generation of vocabularies is now emerging, designed to exploit the electronic media more effectively. A good understanding of the previous generation is still essential for effective access to collections indexed with them. An important object of ISO 25964 as a whole is to support data exchange and other forms of interoperability in circumstances in which more than one structured vocabulary is applied within one retrieval system or network. Sometimes one vocabulary has to be mapped to another, and it is important to understand both the potential and the limitations of such mappings. In other systems, a thesaurus is mapped to a classification scheme, or an ontology to a thesaurus. Comprehensive interoperability needs to cover the whole range of vocabulary types, whether young or old. Concepts in different vocabularies are related only in that they have the same or similar meaning. However, the meaning can be found in a number of different aspects within each particular type of structured vocabulary: - within terms or captions selected in different languages; - in the notation assigned indicating a place within a larger hierarchy; - in the definition, scope notes, history notes and other notes that explain the significance of that concept; and - in explicit relationships to other concepts or entities within the same vocabulary. In order to create mappings from one structured vocabulary to another it is first necessary to understand, within the context of each different type of structured vocabulary, the significance and relative importance of each of the different elements in defining the meaning of that particular concept. ISO 25964-1 describes the key characteristics of thesauri along with additional advice on best practice. ISO 25964-2 focuses on other types of vocabulary and does not attempt to cover all aspects of good practice. It concentrates on those aspects which need to be understood if one of the vocabularies is to work effectively alongside one or more of the others. Recognizing that a new standard cannot be applied to some existing vocabularies, this part of ISO 25964 provides informative description alongside the recommendations, the aim of which is to enable users and system developers to interpret and implement the existing vocabularies effectively. The remainder of ISO 25964-2 deals with the principles and practicalities of establishing mappings between vocabularies.
  18. Börner, K.: Atlas of knowledge : anyone can map (2015) 0.03
    0.028464012 = product of:
      0.085392036 = sum of:
        0.059414867 = weight(_text_:network in 3355) [ClassicSimilarity], result of:
          0.059414867 = score(doc=3355,freq=2.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.29521978 = fieldWeight in 3355, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.046875 = fieldNorm(doc=3355)
        0.025977168 = product of:
          0.051954336 = sum of:
            0.051954336 = weight(_text_:22 in 3355) [ClassicSimilarity], result of:
              0.051954336 = score(doc=3355,freq=4.0), product of:
                0.1582543 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045191888 = queryNorm
                0.32829654 = fieldWeight in 3355, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3355)
          0.5 = coord(1/2)
      0.33333334 = coord(2/6)
    
    Content
    One of a series of three publications influenced by the travelling exhibit Places & Spaces: Mapping Science, curated by the Cyberinfrastructure for Network Science Center at Indiana University. - Additional materials can be found at http://http://scimaps.org/atlas2. Erweitert durch: Börner, Katy. Atlas of Science: Visualizing What We Know.
    Date
    22. 1.2017 16:54:03
    22. 1.2017 17:10:56
  19. Curras, E.: Ontologies, taxonomy and thesauri in information organisation and retrieval (2010) 0.03
    0.027618185 = product of:
      0.082854554 = sum of:
        0.033342157 = weight(_text_:computer in 3276) [ClassicSimilarity], result of:
          0.033342157 = score(doc=3276,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.20188503 = fieldWeight in 3276, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3276)
        0.049512394 = weight(_text_:network in 3276) [ClassicSimilarity], result of:
          0.049512394 = score(doc=3276,freq=2.0), product of:
            0.2012564 = queryWeight, product of:
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.045191888 = queryNorm
            0.2460165 = fieldWeight in 3276, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4533744 = idf(docFreq=1398, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3276)
      0.33333334 = coord(2/6)
    
    Abstract
    The originality of this book, which deals with such a new subject matter, lies in the application of methods and concepts never used before - such as Ontologies and Taxonomies, as well as Thesauri - to the ordering of knowledge based on primary information. Chapters in the book also examine the study of Ontologies, Taxonomies and Thesauri from the perspective of Systematics and General Systems Theory. "Ontologies, Taxonomy and Thesauri in Information Organisation and Retrieval" will be extremely useful to those operating within the network of related fields, which includes Documentation and Information Science.
    Content
    Inhalt: 1. From classifications to ontologies Knowledge - A new concept of knowledge - Knowledge and information - Knowledge organisation - Knowledge organisation and representation - Cognitive sciences - Talent management - Learning systematisation - Historical evolution - From classification to knowledge organisation - Why ontologies exist - Ontologies - The structure of ontologies 2. Taxonomies and thesauri From ordering to taxonomy - The origins of taxonomy - Hierarchical and horizontal order - Correlation with classifications - Taxonomy in computer science - Computing taxonomy - Definitions - Virtual taxonomy, cybernetic taxonomy - Taxonomy in Information Science - Similarities between taxonomies and thesauri - ifferences between taxonomies and thesauri 3. Thesauri Terminology in classification systems - Terminological languages - Thesauri - Thesauri definitions - Conditions that a thesaurus must fulfil - Historical evolution - Classes of thesauri 4. Thesauri in (cladist) systematics Systematics - Systematics as a noun - Definitions and historic evolution over time - Differences between taxonomy and systematics - Systematics in thesaurus construction theory - Classic, numerical and cladist systematics - Classic systematics in information science - Numerical systematics in information science - Thesauri in cladist systematics - Systematics in information technology - Some examples 5. Thesauri in systems theory Historical evolution - Approach to systems - Systems theory applied to the construction of thesauri - Components - Classes of system - Peculiarities of these systems - Working methods - Systems theory applied to ontologies and taxonomies
  20. Schutz, A.; Buitelaar, P.: RelExt: a tool for relation extraction from text in ontology extension (2005) 0.03
    0.026977068 = product of:
      0.0809312 = sum of:
        0.047589045 = weight(_text_:services in 1078) [ClassicSimilarity], result of:
          0.047589045 = score(doc=1078,freq=4.0), product of:
            0.16591617 = queryWeight, product of:
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.045191888 = queryNorm
            0.28682584 = fieldWeight in 1078, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.6713707 = idf(docFreq=3057, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1078)
        0.033342157 = weight(_text_:computer in 1078) [ClassicSimilarity], result of:
          0.033342157 = score(doc=1078,freq=2.0), product of:
            0.16515417 = queryWeight, product of:
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.045191888 = queryNorm
            0.20188503 = fieldWeight in 1078, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.6545093 = idf(docFreq=3109, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1078)
      0.33333334 = coord(2/6)
    
    Abstract
    Domain ontologies very rarely model verbs as relations holding between concepts. However, the role of the verb as a central connecting element between concepts is undeniable. Verbs specify the interaction between the participants of some action or event by expressing relations between them. In parallel, it can be argued from an ontology engineering point of view that verbs express a relation between two classes that specify domain and range. The work described here is concerned with relation extraction for ontology extension along these lines. We describe a system (RelExt) that is capable of automatically identifying highly relevant triples (pairs of concepts connected by a relation) over concepts from an existing ontology. RelExt works by extracting relevant verbs and their grammatical arguments (i.e. terms) from a domain-specific text collection and computing corresponding relations through a combination of linguistic and statistical processing. The paper includes a detailed description of the system architecture and evaluation results on a constructed benchmark. RelExt has been developed in the context of the SmartWeb project, which aims at providing intelligent information services via mobile broadband devices on the FIFA World Cup that will be hosted in Germany in 2006. Such services include location based navigational information as well as question answering in the football domain.
    Series
    Lecture notes in computer science; vol. xxx

Years

Languages

  • e 209
  • d 24

Types

  • a 168
  • el 60
  • m 23
  • x 13
  • s 10
  • n 7
  • p 2
  • r 1
  • More… Less…

Subjects

Classifications