Search (62 results, page 1 of 4)

  • × theme_ss:"Computerlinguistik"
  • × year_i:[2000 TO 2010}
  1. Green, R.: Automated identification of frame semantic relational structures (2000) 0.11
    0.11243666 = product of:
      0.14991555 = sum of:
        0.04016114 = weight(_text_:c in 110) [ClassicSimilarity], result of:
          0.04016114 = score(doc=110,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.2667763 = fieldWeight in 110, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.0546875 = fieldNorm(doc=110)
        0.07430784 = weight(_text_:et in 110) [ClassicSimilarity], result of:
          0.07430784 = score(doc=110,freq=2.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.36287874 = fieldWeight in 110, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=110)
        0.035446577 = product of:
          0.07089315 = sum of:
            0.07089315 = weight(_text_:al in 110) [ClassicSimilarity], result of:
              0.07089315 = score(doc=110,freq=2.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.35444298 = fieldWeight in 110, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=110)
          0.5 = coord(1/2)
      0.75 = coord(3/4)
    
    Source
    Dynamism and stability in knowledge organization: Proceedings of the 6th International ISKO-Conference, 10-13 July 2000, Toronto, Canada. Ed.: C. Beghtol et al
  2. He, Q.: ¬A study of the strength indexes in co-word analysis (2000) 0.10
    0.09637429 = product of:
      0.12849905 = sum of:
        0.03442384 = weight(_text_:c in 111) [ClassicSimilarity], result of:
          0.03442384 = score(doc=111,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.22866541 = fieldWeight in 111, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.046875 = fieldNorm(doc=111)
        0.06369243 = weight(_text_:et in 111) [ClassicSimilarity], result of:
          0.06369243 = score(doc=111,freq=2.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.3110389 = fieldWeight in 111, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.046875 = fieldNorm(doc=111)
        0.03038278 = product of:
          0.06076556 = sum of:
            0.06076556 = weight(_text_:al in 111) [ClassicSimilarity], result of:
              0.06076556 = score(doc=111,freq=2.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.30380827 = fieldWeight in 111, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.046875 = fieldNorm(doc=111)
          0.5 = coord(1/2)
      0.75 = coord(3/4)
    
    Source
    Dynamism and stability in knowledge organization: Proceedings of the 6th International ISKO-Conference, 10-13 July 2000, Toronto, Canada. Ed.: C. Beghtol et al
  3. Mustafa el Hadi, W.: Dynamics of the linguistic paradigm in information retrieval (2000) 0.10
    0.09637429 = product of:
      0.12849905 = sum of:
        0.03442384 = weight(_text_:c in 151) [ClassicSimilarity], result of:
          0.03442384 = score(doc=151,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.22866541 = fieldWeight in 151, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.046875 = fieldNorm(doc=151)
        0.06369243 = weight(_text_:et in 151) [ClassicSimilarity], result of:
          0.06369243 = score(doc=151,freq=2.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.3110389 = fieldWeight in 151, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.046875 = fieldNorm(doc=151)
        0.03038278 = product of:
          0.06076556 = sum of:
            0.06076556 = weight(_text_:al in 151) [ClassicSimilarity], result of:
              0.06076556 = score(doc=151,freq=2.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.30380827 = fieldWeight in 151, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.046875 = fieldNorm(doc=151)
          0.5 = coord(1/2)
      0.75 = coord(3/4)
    
    Source
    Dynamism and stability in knowledge organization: Proceedings of the 6th International ISKO-Conference, 10-13 July 2000, Toronto, Canada. Ed.: C. Beghtol et al
  4. Ou, S.; Khoo, C.; Goh, D.H.; Heng, H.-Y.: Automatic discourse parsing of sociology dissertation abstracts as sentence categorization (2004) 0.06
    0.06424952 = product of:
      0.08566603 = sum of:
        0.022949224 = weight(_text_:c in 2676) [ClassicSimilarity], result of:
          0.022949224 = score(doc=2676,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.1524436 = fieldWeight in 2676, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.03125 = fieldNorm(doc=2676)
        0.042461623 = weight(_text_:et in 2676) [ClassicSimilarity], result of:
          0.042461623 = score(doc=2676,freq=2.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.20735928 = fieldWeight in 2676, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.03125 = fieldNorm(doc=2676)
        0.020255188 = product of:
          0.040510375 = sum of:
            0.040510375 = weight(_text_:al in 2676) [ClassicSimilarity], result of:
              0.040510375 = score(doc=2676,freq=2.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.20253885 = fieldWeight in 2676, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2676)
          0.5 = coord(1/2)
      0.75 = coord(3/4)
    
    Content
    1. Introduction This paper reports our initial effort to develop an automatic method for parsing the discourse structure of sociology dissertation abstracts. This study is part of a broader study to develop a method for multi-document summarization. Accurate discourse parsing will make it easier to perform automatic multi-document summarization of dissertation abstracts. In a previous study, we determined that the macro-level structure of dissertation abstracts typically has five sections (Khoo et al., 2002). In this study, we treated discourse parsing as a text categorization problem - assigning each sentence in a dissertation abstract to one of the five predefined sections or categories. Decision tree induction, a machine-learning method, was applied to word tokens found in the abstracts to construct a decision tree model for the categorization purpose. Decision tree induction was selected primarily because decision tree models are easy to interpret and can be converted to rules that can be incorporated in other computer programs. A well-known decision-tree induction program, C5.0 (Quinlan, 1993), was used in this study.
  5. WordHoard: finding multiword units (20??) 0.05
    0.054877207 = product of:
      0.10975441 = sum of:
        0.07430784 = weight(_text_:et in 1123) [ClassicSimilarity], result of:
          0.07430784 = score(doc=1123,freq=2.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.36287874 = fieldWeight in 1123, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1123)
        0.035446577 = product of:
          0.07089315 = sum of:
            0.07089315 = weight(_text_:al in 1123) [ClassicSimilarity], result of:
              0.07089315 = score(doc=1123,freq=2.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.35444298 = fieldWeight in 1123, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1123)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    WordHoard defines a multiword unit as a special type of collocate in which the component words comprise a meaningful phrase. For example, "Knight of the Round Table" is a meaningful multiword unit or phrase. WordHoard uses the notion of a pseudo-bigram to generalize the computation of bigram (two word) statistical measures to phrases (n-grams) longer than two words, and to allow comparisons of these measures for phrases with different word counts. WordHoard applies the localmaxs algorithm of Silva et al. to the pseudo-bigrams to identify potential compositional phrases that "stand out" in a text. WordHoard can also filter two and three word phrases using the word class filters suggested by Justeson and Katz.
  6. Nait-Baha, L.; Jackiewicz, A.; Djioua, B.; Laublet, P.: Query reformulation for information retrieval on the Web using the point of view methodology : preliminary results (2001) 0.05
    0.047037605 = product of:
      0.09407521 = sum of:
        0.06369243 = weight(_text_:et in 249) [ClassicSimilarity], result of:
          0.06369243 = score(doc=249,freq=2.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.3110389 = fieldWeight in 249, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.046875 = fieldNorm(doc=249)
        0.03038278 = product of:
          0.06076556 = sum of:
            0.06076556 = weight(_text_:al in 249) [ClassicSimilarity], result of:
              0.06076556 = score(doc=249,freq=2.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.30380827 = fieldWeight in 249, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.046875 = fieldNorm(doc=249)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    The work we are presenting is devoted to the information collected on the WWW. By the term collected we mean the whole process of retrieving, extracting and presenting results to the user. This research is part of the RAP (Research, Analyze, Propose) project in which we propose to combine two methods: (i) query reformulation using linguistic markers according to a given point of view; and (ii) text semantic analysis by means of contextual exploration results (Descles, 1991). The general project architecture describing the interactions between the users, the RAP system and the WWW search engines is presented in Nait-Baha et al. (1998). We will focus this paper on showing how we use linguistic markers to reformulate the queries according to a given point of view
  7. Schmidt, R.: Maschinelle Text-Ton-Synchronisation in Wissenschaft und Wirtschaft (2000) 0.04
    0.039198004 = product of:
      0.07839601 = sum of:
        0.053077027 = weight(_text_:et in 5559) [ClassicSimilarity], result of:
          0.053077027 = score(doc=5559,freq=2.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.2591991 = fieldWeight in 5559, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5559)
        0.025318984 = product of:
          0.050637968 = sum of:
            0.050637968 = weight(_text_:al in 5559) [ClassicSimilarity], result of:
              0.050637968 = score(doc=5559,freq=2.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.25317356 = fieldWeight in 5559, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5559)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Tonmaterial in Form von Audio- oder Videoaufnahmen spielt in Bereichen der Wissenschaft, die sich mit verbaler Interaktion beschäftigen, eine bedeutende Rolle. Solche Gebiete sind u,a. die Linguistik, Psychologie, Soziologie und Kriminalistik. Gegenstand der Untersuchung können dabei z.B. die Formen des sprachlichen Handelns und der Sprachvariation in Abhängigkeit von der Situation oder die Ausprägung und Entwicklung von Sprachunterschieden vor dem sozialen Hintergrund sein. Im Rahmen der Analyse eines Gesprächsverlaufs kann beispielsweise die Form der Rederechtsicherung von Interesse sein. In diesem Zusammenhang stellen sich Fragen wie z.B. "Wie bringen Gesprächsteilnehrner Gesprächsbeteiligte dazu, ihre Rede zu unterbrechen?" oder "Wie wehren Gesprächsteilnehmer Unterbrechungsversuche voll anderen Teilnehmern ab?". Denkbar ist hier u.a. nach dem Vorkommen von "ausreden lassen" zu suchen, wobei diese beiden Wörter nicht unbedingt nebeneinander auftreten müssen. Bei der Suche nach Stellen an denen ein Gesprächsteilnehmer Ansprüche oder Forderungen an einen Gesprächspartner stellt, können die flektierten Formen der Modalverben wie z.B. "müssen", "sollen" oder "dürfen" für die Anfrage wichtig sein, während Konnektiva wie "aber", "ja aber" oder "doch" auf oppositive Gesprächsabschnitte verweisen können. Näheres zur gesprächsanalytischen Methodik kann Deppermann (1999) und Brünner et al. (1999) entnommen werden. In dem Bereich der Linguistik, die den Gebrauch von gesprochener Sprache in offiziellen und privaten Situationen zum Gegenstand hat, sind u.a. auch Aussprachevarianten von großem Interesse. Von der Untersuchung der Sprachfärbungen erhofft man sich detaillierte Aussagen über die Sprechersituation und die regionale (König (1988)) und soziale Herkunft des Sprechers machen zu können. In der Kriminalistik wirken solche Ergebnisse unterstützend bei der Identifizierung von Personen
  8. Niemi, T.; Jämsen, J.: ¬A query language for discovering semantic associations, part II : sample queries and query evaluation (2007) 0.04
    0.039198004 = product of:
      0.07839601 = sum of:
        0.053077027 = weight(_text_:et in 580) [ClassicSimilarity], result of:
          0.053077027 = score(doc=580,freq=2.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.2591991 = fieldWeight in 580, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=580)
        0.025318984 = product of:
          0.050637968 = sum of:
            0.050637968 = weight(_text_:al in 580) [ClassicSimilarity], result of:
              0.050637968 = score(doc=580,freq=2.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.25317356 = fieldWeight in 580, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=580)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    In our query language introduced in Part I (Journal of the American Society for Information Science and Technology. 58(2007) no.11, S.1559-1568) the user can formulate queries to find out (possibly complex) semantic relationships among entities. In this article we demonstrate the usage of our query language and discuss the new applications that it supports. We categorize several query types and give sample queries. The query types are categorized based on whether the entities specified in a query are known or unknown to the user in advance, and whether text information in documents is utilized. Natural language is used to represent the results of queries in order to facilitate correct interpretation by the user. We discuss briefly the issues related to the prototype implementation of the query language and show that an independent operation like Rho (Sheth et al., 2005; Anyanwu & Sheth, 2002, 2003), which presupposes entities of interest to be known in advance, is exceedingly inefficient in emulating the behavior of our query language. The discussion also covers potential problems, and challenges for future work.
  9. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.03
    0.034863282 = product of:
      0.069726564 = sum of:
        0.05198749 = product of:
          0.20794997 = sum of:
            0.20794997 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
              0.20794997 = score(doc=562,freq=2.0), product of:
                0.37000585 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.043643 = queryNorm
                0.56201804 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.25 = coord(1/4)
        0.017739072 = product of:
          0.035478145 = sum of:
            0.035478145 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
              0.035478145 = score(doc=562,freq=2.0), product of:
                0.15283036 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043643 = queryNorm
                0.23214069 = fieldWeight in 562, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=562)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  10. Thiel, M.: Bedingt wahrscheinliche Syntaxbäume (2006) 0.03
    0.031358406 = product of:
      0.06271681 = sum of:
        0.042461623 = weight(_text_:et in 6069) [ClassicSimilarity], result of:
          0.042461623 = score(doc=6069,freq=2.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.20735928 = fieldWeight in 6069, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.03125 = fieldNorm(doc=6069)
        0.020255188 = product of:
          0.040510375 = sum of:
            0.040510375 = weight(_text_:al in 6069) [ClassicSimilarity], result of:
              0.040510375 = score(doc=6069,freq=2.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.20253885 = fieldWeight in 6069, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.03125 = fieldNorm(doc=6069)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Die Tendenz ist eindeutig: wo immer es sinnvoll ist, werden hart' programmierte Lösungen durch Ansätze des Softcomputing ersetzt. Vor allem technische und kommerzielle Bereiche profitieren davon. So finden wir Kransteuerungen und viele andere Anwendungen mit Fuzzy Expertensystemen sowie Bilderkennungssysteme und Entscheidungen über die Kreditvergabe mit Neuronalen Netzen oder auch Methoden des Maschinellen Lernens (vgl. Jafar-Shaghaghi 1994). Ein Prinzip dieser Ansätze ist, dass die Software sich automatisch an die spezielle Situation und Datengrundlage der Anwendung anpasst. Flexibilität der Anpassung und die Fähigkeit zur Verallgemeinerung auf bislang ungesehene Fälle sind implizit in den Methoden vorhanden. Gerade dies ist auch ein typisches Problem, das bei der Beschreibung und vor allem beim Parsen natürlicher Sprache auftritt. Bei der Verarbeitung natürlicher Sprache kommt das leidige Problem der Ambiguität auf verschiedenen Ebenen hinzu. Alternative Regeln schließen sich in ihrer Anwendung in einem Satz meistens gegenseitig aus und sind nicht alle an der aktuellen Stelle gleich wahrscheinlich. Auf diese Problematik wurde schon früh hingewiesen (Thiel 1987, 137 ff.), wo versucht wurde, mit Gewichtungen die Wahrscheinlichkeit von Regeln, Syntaxbäumen, Kategorien und Wortsemantik in den Griff zu bekommen. Das Gewicht eines Syntaxbaumes kann z.B. einfach zugewiesen werden oder berechnet werden als Funktion des Baumes, aus dem er abgeleitet wird, und der angewandten Regel. Ein solches Verfahren wird (Thiel 1987, 152) am Beispiel einer Heuristik für die Inferenzmaschine eines Expertensystems gezeigt. Aber auch bereits in einer sehr frühen Veröffentlichung zur Analyse natürlicher Sprache, an der Zimmermann maßgeblich beteiligt war, wurde auf Vorkommenswahrscheinlichkeiten hingewiesen: "Statistische Auswertung von Typen des Satzbaus, Bau nominaler und verbaler Gruppen ..." (Eggers et al. 1969, 18). Derzeit konzentrieren sich die Ansätze von Vagheit in der Verarbeitung von natürlicher Sprache vor allem auf die Filterung von Texten z.B. in Spam-Filtern und auf probabilistische kontextfreie Grammatiken.
  11. Jones, I.; Cunliffe, D.; Tudhope, D.: Natural language processing and knowledge organization systems as an aid to retrieval (2004) 0.03
    0.027438603 = product of:
      0.054877207 = sum of:
        0.03715392 = weight(_text_:et in 2677) [ClassicSimilarity], result of:
          0.03715392 = score(doc=2677,freq=2.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.18143937 = fieldWeight in 2677, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.02734375 = fieldNorm(doc=2677)
        0.017723288 = product of:
          0.035446577 = sum of:
            0.035446577 = weight(_text_:al in 2677) [ClassicSimilarity], result of:
              0.035446577 = score(doc=2677,freq=2.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.17722149 = fieldWeight in 2677, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=2677)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Content
    1. Introduction The need for research into the application of linguistic techniques in Information Retrieval (IR) in general, and a similar need in faceted Knowledge Organization Systems (KOS) has been indicated by various authors. Smeaton (1997) points out the inherent limitations of conventional approaches to IR based an "bags of words", mainly difficulties caused by lexical ambiguity in the words concerned, and goes an to suggest the possibility of using Natural Language Processing (NLP) in query formulation. Past experience with a faceted retrieval system highlighted the need for integrating the linguistic perspective in order to fully utilise the potential of a KOS (Tudhope et al." 2002). The present research seeks to address some of these needs in using NLP to improve the efficacy of KOS tools in query and retrieval systems. Syntactic parsing and part-of-speech tagging can substantially reduce lexical ambiguity through homograph disambiguation. Given the two strings "1 fable the motion" and "I put the motion an the fable", for instance, the parser used in this research clearly indicates that 'fable' in the first string is a verb, while 'table' in the second string is a noun, a distinction that would be missed in the "bag of words" approach. This syntactic disambiguation enables a more precise matching from free text to the controlled vocabulary of a KOS and vice versa. The use of a general linguistic resource, namely Roget's Thesaurus of English Words and Phrases (RTEWP), as an intermediary in this process, is investigated. The adaptation of the Link parser (Sleator & Temperley, 1993) to the purposes of the research is reported. The design and implementation of the early practical stages of the project are described, and the results of the initial experiments are presented and evaluated. Applications of the techniques developed are foreseen in the areas of query disambiguation, information retrieval and automatic indexing. In the first section of the paper a brief review of the literature and relevant current work in the field is presented. The second section includes reports an the development of algorithms, the construction of data sets and theoretical and experimental work undertaken to date. The third section evaluates the results obtained, and outlines directions for future research.
  12. Semantic role universals and argument linking : theoretical, typological, and psycholinguistic perspectives (2006) 0.03
    0.026002327 = product of:
      0.10400931 = sum of:
        0.10400931 = weight(_text_:et in 3670) [ClassicSimilarity], result of:
          0.10400931 = score(doc=3670,freq=12.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.50792444 = fieldWeight in 3670, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.03125 = fieldNorm(doc=3670)
      0.25 = coord(1/4)
    
    Classification
    ET 750
    ET 455
    ET 400
    RVK
    ET 750
    ET 455
    ET 400
  13. Barriere, C.: Building a concept hierarchy from corpus analysis (2004) 0.02
    0.022949224 = product of:
      0.0917969 = sum of:
        0.0917969 = weight(_text_:c in 6787) [ClassicSimilarity], result of:
          0.0917969 = score(doc=6787,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.6097744 = fieldWeight in 6787, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.125 = fieldNorm(doc=6787)
      0.25 = coord(1/4)
    
  14. Kunze, C.; Wagner, A.: Anwendungsperspektive des GermaNet, eines lexikalisch-semantischen Netzes für das Deutsche (2001) 0.02
    0.01721192 = product of:
      0.06884768 = sum of:
        0.06884768 = weight(_text_:c in 7456) [ClassicSimilarity], result of:
          0.06884768 = score(doc=7456,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.45733082 = fieldWeight in 7456, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.09375 = fieldNorm(doc=7456)
      0.25 = coord(1/4)
    
  15. Hull, D.; Ait-Mokhtar, S.; Chuat, M.; Eisele, A.; Gaussier, E.; Grefenstette, G.; Isabelle, P.; Samulesson, C.; Segand, F.: Language technologies and patent search and classification (2001) 0.02
    0.01721192 = product of:
      0.06884768 = sum of:
        0.06884768 = weight(_text_:c in 6318) [ClassicSimilarity], result of:
          0.06884768 = score(doc=6318,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.45733082 = fieldWeight in 6318, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.09375 = fieldNorm(doc=6318)
      0.25 = coord(1/4)
    
  16. Conceptual structures : logical, linguistic, and computational issues. 8th International Conference on Conceptual Structures, ICCS 2000, Darmstadt, Germany, August 14-18, 2000 (2000) 0.02
    0.016201654 = product of:
      0.03240331 = sum of:
        0.01721192 = weight(_text_:c in 691) [ClassicSimilarity], result of:
          0.01721192 = score(doc=691,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.114332706 = fieldWeight in 691, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.0234375 = fieldNorm(doc=691)
        0.01519139 = product of:
          0.03038278 = sum of:
            0.03038278 = weight(_text_:al in 691) [ClassicSimilarity], result of:
              0.03038278 = score(doc=691,freq=2.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.15190414 = fieldWeight in 691, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=691)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Computer scientists create models of a perceived reality. Through AI techniques, these models aim at providing the basic support for emulating cognitive behavior such as reasoning and learning, which is one of the main goals of the Al research effort. Such computer models are formed through the interaction of various acquisition and inference mechanisms: perception, concept learning, conceptual clustering, hypothesis testing, probabilistic inference, etc., and are represented using different paradigms tightly linked to the processes that use them. Among these paradigms let us cite: biological models (neural nets, genetic programming), logic-based models (first-order logic, modal logic, rule-based systems), virtual reality models (object systems, agent systems), probabilistic models (Bayesian nets, fuzzy logic), linguistic models (conceptual dependency graphs, language-based rep resentations), etc. One of the strengths of the Conceptual Graph (CG) theory is its versatility in terms of the representation paradigms under which it falls. It can be viewed and therefore used, under different representation paradigms, which makes it a popular choice for a wealth of applications. Its full coupling with different cognitive processes lead to the opening of the field toward related research communities such as the Description Logic, Formal Concept Analysis, and Computational Linguistic communities. We now see more and more research results from one community enrich the other, laying the foundations of common philosophical grounds from which a successful synergy can emerge. ICCS 2000 embodies this spirit of research collaboration. It presents a set of papers that we believe, by their exposure, will benefit the whole community. For instance, the technical program proposes tracks on Conceptual Ontologies, Language, Formal Concept Analysis, Computational Aspects of Conceptual Structures, and Formal Semantics, with some papers on pragmatism and human related aspects of computing. Never before was the program of ICCS formed by so heterogeneously rooted theories of knowledge representation and use. We hope that this swirl of ideas will benefit you as much as it already has benefited us while putting together this program
    Content
    Concepts and Language: The Role of Conceptual Structure in Human Evolution (Keith Devlin) - Concepts in Linguistics - Concepts in Natural Language (Gisela Harras) - Patterns, Schemata, and Types: Author Support through Formalized Experience (Felix H. Gatzemeier) - Conventions and Notations for Knowledge Representation and Retrieval (Philippe Martin) - Conceptual Ontology: Ontology, Metadata, and Semiotics (John F. Sowa) - Pragmatically Yours (Mary Keeler) - Conceptual Modeling for Distributed Ontology Environments (Deborah L. McGuinness) - Discovery of Class Relations in Exception Structured Knowledge Bases (Hendra Suryanto, Paul Compton) - Conceptual Graphs: Perspectives: CGs Applications: Where Are We 7 Years after the First ICCS ? (Michel Chein, David Genest) - The Engineering of a CC-Based System: Fundamental Issues (Guy W. Mineau) - Conceptual Graphs, Metamodeling, and Notation of Concepts (Olivier Gerbé, Guy W. Mineau, Rudolf K. Keller) - Knowledge Representation and Reasonings: Based on Graph Homomorphism (Marie-Laure Mugnier) - User Modeling Using Conceptual Graphs for Intelligent Agents (James F. Baldwin, Trevor P. Martin, Aimilia Tzanavari) - Towards a Unified Querying System of Both Structured and Semi-structured Imprecise Data Using Fuzzy View (Patrice Buche, Ollivier Haemmerlé) - Formal Semantics of Conceptual Structures: The Extensional Semantics of the Conceptual Graph Formalism (Guy W. Mineau) - Semantics of Attribute Relations in Conceptual Graphs (Pavel Kocura) - Nested Concept Graphs and Triadic Power Context Families (Susanne Prediger) - Negations in Simple Concept Graphs (Frithjof Dau) - Extending the CG Model by Simulations (Jean-François Baget) - Contextual Logic and Formal Concept Analysis: Building and Structuring Description Logic Knowledge Bases: Using Least Common Subsumers and Concept Analysis (Franz Baader, Ralf Molitor) - On the Contextual Logic of Ordinal Data (Silke Pollandt, Rudolf Wille) - Boolean Concept Logic (Rudolf Wille) - Lattices of Triadic Concept Graphs (Bernd Groh, Rudolf Wille) - Formalizing Hypotheses with Concepts (Bernhard Ganter, Sergei 0. Kuznetsov) - Generalized Formal Concept Analysis (Laurent Chaudron, Nicolas Maille) - A Logical Generalization of Formal Concept Analysis (Sébastien Ferré, Olivier Ridoux) - On the Treatment of Incomplete Knowledge in Formal Concept Analysis (Peter Burmeister, Richard Holzer) - Conceptual Structures in Practice: Logic-Based Networks: Concept Graphs and Conceptual Structures (Peter W. Eklund) - Conceptual Knowledge Discovery and Data Analysis (Joachim Hereth, Gerd Stumme, Rudolf Wille, Uta Wille) - CEM - A Conceptual Email Manager (Richard Cole, Gerd Stumme) - A Contextual-Logic Extension of TOSCANA (Peter Eklund, Bernd Groh, Gerd Stumme, Rudolf Wille) - A Conceptual Graph Model for W3C Resource Description Framework (Olivier Corby, Rose Dieng, Cédric Hébert) - Computational Aspects of Conceptual Structures: Computing with Conceptual Structures (Bernhard Ganter) - Symmetry and the Computation of Conceptual Structures (Robert Levinson) An Introduction to SNePS 3 (Stuart C. Shapiro) - Composition Norm Dynamics Calculation with Conceptual Graphs (Aldo de Moor) - From PROLOG++ to PROLOG+CG: A CG Object-Oriented Logic Programming Language (Adil Kabbaj, Martin Janta-Polczynski) - A Cost-Bounded Algorithm to Control Events Generalization (Gaël de Chalendar, Brigitte Grau, Olivier Ferret)
  17. ¬The semantics of relationships : an interdisciplinary perspective (2002) 0.02
    0.016036259 = product of:
      0.064145036 = sum of:
        0.064145036 = weight(_text_:c in 1430) [ClassicSimilarity], result of:
          0.064145036 = score(doc=1430,freq=10.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.42609286 = fieldWeight in 1430, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1430)
      0.25 = coord(1/4)
    
    Content
    Enthält die Beiträge: Pt.1: Types of relationships: CRUDE, D.A.: Hyponymy and its varieties; FELLBAUM, C.: On the semantics of troponymy; PRIBBENOW, S.: Meronymic relationships: from classical mereology to complex part-whole relations; KHOO, C. u.a.: The many facets of cause-effect relation - Pt.2: Relationships in knowledge representation and reasoning: GREEN, R.: Internally-structured conceptual models in cognitive semantics; HOVY, E.: Comparing sets of semantic relations in ontologies; GUARINO, N., C. WELTY: Identity and subsumption; JOUIS; C.: Logic of relationships - Pt.3: Applications of relationships: EVENS, M.: Thesaural relations in information retrieval; KHOO, C., S.H. MYAENG: Identifying semantic relations in text for information retrieval and information extraction; McCRAY, A.T., O. BODENREICHER: A conceptual framework for the biiomedical domain; HETZLER, B.: Visual analysis and exploration of relationships
  18. Melzer, C.: ¬Der Maschine anpassen : PC-Spracherkennung - Programme sind mittlerweile alltagsreif (2005) 0.02
    0.0152141815 = product of:
      0.030428363 = sum of:
        0.02008057 = weight(_text_:c in 4044) [ClassicSimilarity], result of:
          0.02008057 = score(doc=4044,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.13338815 = fieldWeight in 4044, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4044)
        0.010347793 = product of:
          0.020695586 = sum of:
            0.020695586 = weight(_text_:22 in 4044) [ClassicSimilarity], result of:
              0.020695586 = score(doc=4044,freq=2.0), product of:
                0.15283036 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043643 = queryNorm
                0.1354154 = fieldWeight in 4044, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=4044)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Date
    3. 5.1997 8:44:22
  19. Sidhom, S.; Hassoun, M.: Morpho-syntactic parsing to text mining environment : NP recognition model to knowledge visualization and information (2003) 0.01
    0.014343265 = product of:
      0.05737306 = sum of:
        0.05737306 = weight(_text_:c in 3546) [ClassicSimilarity], result of:
          0.05737306 = score(doc=3546,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.381109 = fieldWeight in 3546, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.078125 = fieldNorm(doc=3546)
      0.25 = coord(1/4)
    
    Source
    Tendencias de investigación en organización del conocimient: IV Cologuio International de Ciencas de la Documentación , VI Congreso del Capitulo Espanol de ISKO = Trends in knowledge organization research. Eds.: J.A. Frias u. C. Travieso
  20. Liu, S.; Liu, F.; Yu, C.; Meng, W.: ¬An effective approach to document retrieval via utilizing WordNet and recognizing phrases (2004) 0.01
    0.014343265 = product of:
      0.05737306 = sum of:
        0.05737306 = weight(_text_:c in 4078) [ClassicSimilarity], result of:
          0.05737306 = score(doc=4078,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.381109 = fieldWeight in 4078, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.078125 = fieldNorm(doc=4078)
      0.25 = coord(1/4)
    

Languages

  • e 40
  • d 22

Types

  • a 52
  • m 7
  • el 4
  • s 4
  • x 2
  • More… Less…