Search (33 results, page 2 of 2)

  • × theme_ss:"Computerlinguistik"
  • × year_i:[2010 TO 2020}
  1. Ramisch, C.: Multiword expressions acquisition : a generic and open framework (2015) 0.00
    0.003881862 = product of:
      0.011645585 = sum of:
        0.011645585 = product of:
          0.034936756 = sum of:
            0.034936756 = weight(_text_:k in 1649) [ClassicSimilarity], result of:
              0.034936756 = score(doc=1649,freq=4.0), product of:
                0.15658903 = queryWeight, product of:
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0438652 = queryNorm
                0.22311112 = fieldWeight in 1649, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1649)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Classification
    BFP (FH K)
    GHBS
    BFP (FH K)
  2. Sankarasubramaniam, Y.; Ramanathan, K.; Ghosh, S.: Text summarization using Wikipedia (2014) 0.00
    0.0034311134 = product of:
      0.01029334 = sum of:
        0.01029334 = product of:
          0.03088002 = sum of:
            0.03088002 = weight(_text_:k in 2693) [ClassicSimilarity], result of:
              0.03088002 = score(doc=2693,freq=2.0), product of:
                0.15658903 = queryWeight, product of:
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0438652 = queryNorm
                0.19720423 = fieldWeight in 2693, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2693)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
  3. Savoy, J.: Text representation strategies : an example with the State of the union addresses (2016) 0.00
    0.0034311134 = product of:
      0.01029334 = sum of:
        0.01029334 = product of:
          0.03088002 = sum of:
            0.03088002 = weight(_text_:k in 3042) [ClassicSimilarity], result of:
              0.03088002 = score(doc=3042,freq=2.0), product of:
                0.15658903 = queryWeight, product of:
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0438652 = queryNorm
                0.19720423 = fieldWeight in 3042, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3042)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    Based on State of the Union addresses from 1790 to 2014 (225 speeches delivered by 42 presidents), this paper describes and evaluates different text representation strategies. To determine the most important words of a given text, the term frequencies (tf) or the tf?idf weighting scheme can be applied. Recently, latent Dirichlet allocation (LDA) has been proposed to define the topics included in a corpus. As another strategy, this study proposes to apply a vocabulary specificity measure (Z?score) to determine the most significantly overused word-types or short sequences of them. Our experiments show that the simple term frequency measure is not able to discriminate between specific terms associated with a document or a set of texts. Using the tf idf or LDA approach, the selection requires some arbitrary decisions. Based on the term-specific measure (Z?score), the term selection has a clear theoretical basis. Moreover, the most significant sentences for each presidency can be determined. As another facet, we can visualize the dynamic evolution of usage of some terms associated with their specificity measures. Finally, this technique can be employed to define the most important lexical leaders introducing terms overused by the k following presidencies.
  4. Lian, T.; Yu, C.; Wang, W.; Yuan, Q.; Hou, Z.: Doctoral dissertations on tourism in China : a co-word analysis (2016) 0.00
    0.0034311134 = product of:
      0.01029334 = sum of:
        0.01029334 = product of:
          0.03088002 = sum of:
            0.03088002 = weight(_text_:k in 3178) [ClassicSimilarity], result of:
              0.03088002 = score(doc=3178,freq=2.0), product of:
                0.15658903 = queryWeight, product of:
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0438652 = queryNorm
                0.19720423 = fieldWeight in 3178, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3178)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Abstract
    The aim of this paper is to map the foci of research in doctoral dissertations on tourism in China. In the paper, coword analysis is applied, with keywords coming from six public dissertation databases, i.e. CDFD, Wanfang Data, NLC, CALIS, ISTIC, and NSTL, as well as some university libraries providing doctoral dissertations on tourism. Altogether we have examined 928 doctoral dissertations on tourism written between 1989 and 2013. Doctoral dissertations on tourism in China involve 36 first level disciplines and 102 secondary level disciplines. We collect the top 68 keywords of practical significance in tourism which are mentioned at least four times or more. These keywords are classified into 12 categories based on co-word analysis, including cluster analysis, strategic diagrams analysis, and social network analysis. According to the strategic diagram of the 12 categories, we find the mature and immature areas in tourism study. From social networks, we can see the social network maps of original co-occurrence matrix and k-cores analysis of binary matrix. The paper provides valuable insight into the study of tourism by analyzing doctoral dissertations on tourism in China.
  5. Lhadj, L.S.; Boughanem, M.; Amrouche, K.: Enhancing information retrieval through concept-based language modeling and semantic smoothing (2016) 0.00
    0.0034311134 = product of:
      0.01029334 = sum of:
        0.01029334 = product of:
          0.03088002 = sum of:
            0.03088002 = weight(_text_:k in 3221) [ClassicSimilarity], result of:
              0.03088002 = score(doc=3221,freq=2.0), product of:
                0.15658903 = queryWeight, product of:
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0438652 = queryNorm
                0.19720423 = fieldWeight in 3221, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3221)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
  6. Järvelin, A.; Keskustalo, H.; Sormunen, E.; Saastamoinen, M.; Kettunen, K.: Information retrieval from historical newspaper collections in highly inflectional languages : a query expansion approach (2016) 0.00
    0.0034311134 = product of:
      0.01029334 = sum of:
        0.01029334 = product of:
          0.03088002 = sum of:
            0.03088002 = weight(_text_:k in 3223) [ClassicSimilarity], result of:
              0.03088002 = score(doc=3223,freq=2.0), product of:
                0.15658903 = queryWeight, product of:
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0438652 = queryNorm
                0.19720423 = fieldWeight in 3223, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3223)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
  7. K., Vani; Gupta, D.: Unmasking text plagiarism using syntactic-semantic based natural language processing techniques : comparisons, analysis and challenges (2018) 0.00
    0.0034311134 = product of:
      0.01029334 = sum of:
        0.01029334 = product of:
          0.03088002 = sum of:
            0.03088002 = weight(_text_:k in 5084) [ClassicSimilarity], result of:
              0.03088002 = score(doc=5084,freq=2.0), product of:
                0.15658903 = queryWeight, product of:
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0438652 = queryNorm
                0.19720423 = fieldWeight in 5084, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5084)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
  8. Gill, A.J.; Hinrichs-Krapels, S.; Blanke, T.; Grant, J.; Hedges, M.; Tanner, S.: Insight workflow : systematically combining human and computational methods to explore textual data (2017) 0.00
    0.0033317097 = product of:
      0.009995129 = sum of:
        0.009995129 = product of:
          0.029985385 = sum of:
            0.029985385 = weight(_text_:29 in 3682) [ClassicSimilarity], result of:
              0.029985385 = score(doc=3682,freq=2.0), product of:
                0.15430406 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0438652 = queryNorm
                0.19432661 = fieldWeight in 3682, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3682)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Date
    16.11.2017 14:00:29
  9. Fóris, A.: Network theory and terminology (2013) 0.00
    0.0033017385 = product of:
      0.009905215 = sum of:
        0.009905215 = product of:
          0.029715646 = sum of:
            0.029715646 = weight(_text_:22 in 1365) [ClassicSimilarity], result of:
              0.029715646 = score(doc=1365,freq=2.0), product of:
                0.15360846 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0438652 = queryNorm
                0.19345059 = fieldWeight in 1365, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1365)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Date
    2. 9.2014 21:22:48
  10. RWI/PH: Auf der Suche nach dem entscheidenden Wort : die Häufung bestimmter Wörter innerhalb eines Textes macht diese zu Schlüsselwörtern (2012) 0.00
    0.0029113963 = product of:
      0.008734189 = sum of:
        0.008734189 = product of:
          0.026202565 = sum of:
            0.026202565 = weight(_text_:k in 331) [ClassicSimilarity], result of:
              0.026202565 = score(doc=331,freq=4.0), product of:
                0.15658903 = queryWeight, product of:
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0438652 = queryNorm
                0.16733333 = fieldWeight in 331, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=331)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Content
    "Die Dresdner Wissenschaftler haben die semantischen Eigenschaften von Texten mathematisch untersucht, indem sie zehn verschiedene englische Texte in unterschiedlichen Formen kodierten. Dazu zählt unter anderem die englische Ausgabe von Leo Tolstois "Krieg und Frieden". Beispielsweise übersetzten die Forscher Buchstaben innerhalb eines Textes in eine Binär-Sequenz. Dazu ersetzten sie alle Vokale durch eine Eins und alle Konsonanten durch eine Null. Mit Hilfe weiterer mathematischer Funktionen beleuchteten die Wissenschaftler dabei verschiedene Ebenen des Textes, also sowohl einzelne Vokale, Buchstaben als auch ganze Wörter, die in verschiedenen Formen kodiert wurden. Innerhalb des ganzen Textes lassen sich so wiederkehrende Muster finden. Diesen Zusammenhang innerhalb des Textes bezeichnet man als Langzeitkorrelation. Diese gibt an, ob zwei Buchstaben an beliebig weit voneinander entfernten Textstellen miteinander in Verbindung stehen - beispielsweise gibt es wenn wir an einer Stelle einen Buchstaben "K" finden, eine messbare höhere Wahrscheinlichkeit den Buchstaben "K" einige Seiten später nochmal zu finden. "Es ist zu erwarten, dass wenn es in einem Buch an einer Stelle um Krieg geht, die Wahrscheinlichkeit hoch ist das Wort Krieg auch einige Seiten später zu finden. Überraschend ist es, dass wir die hohe Wahrscheinlichkeit auch auf der Buchstabenebene finden", so Altmann.
  11. Kajanan, S.; Bao, Y.; Datta, A.; VanderMeer, D.; Dutta, K.: Efficient automatic search query formulation using phrase-level analysis (2014) 0.00
    0.0027448907 = product of:
      0.008234672 = sum of:
        0.008234672 = product of:
          0.024704017 = sum of:
            0.024704017 = weight(_text_:k in 1264) [ClassicSimilarity], result of:
              0.024704017 = score(doc=1264,freq=2.0), product of:
                0.15658903 = queryWeight, product of:
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.0438652 = queryNorm
                0.15776339 = fieldWeight in 1264, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.569778 = idf(docFreq=3384, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1264)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
  12. Rötzer, F.: KI-Programm besser als Menschen im Verständnis natürlicher Sprache (2018) 0.00
    0.0026413908 = product of:
      0.007924172 = sum of:
        0.007924172 = product of:
          0.023772515 = sum of:
            0.023772515 = weight(_text_:22 in 4217) [ClassicSimilarity], result of:
              0.023772515 = score(doc=4217,freq=2.0), product of:
                0.15360846 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0438652 = queryNorm
                0.15476047 = fieldWeight in 4217, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4217)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Date
    22. 1.2018 11:32:44
  13. Deventer, J.P. van; Kruger, C.J.; Johnson, R.D.: Delineating knowledge management through lexical analysis : a retrospective (2015) 0.00
    0.0023112171 = product of:
      0.006933651 = sum of:
        0.006933651 = product of:
          0.020800952 = sum of:
            0.020800952 = weight(_text_:22 in 3807) [ClassicSimilarity], result of:
              0.020800952 = score(doc=3807,freq=2.0), product of:
                0.15360846 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0438652 = queryNorm
                0.1354154 = fieldWeight in 3807, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=3807)
          0.33333334 = coord(1/3)
      0.33333334 = coord(1/3)
    
    Date
    20. 1.2015 18:30:22

Languages

  • e 26
  • d 7

Types

Classifications