Search (459 results, page 22 of 23)

  • × theme_ss:"Computerlinguistik"
  1. Lund, B.D.; Wang, T.; Mannuru, N.R.; Nie, B.; Shimray, S.; Wang, Z.: ChatGPT and a new academic reality : artificial Intelligence-written research papers and the ethics of the large language models in scholarly publishing (2023) 0.00
    0.0035296201 = product of:
      0.01058886 = sum of:
        0.01058886 = weight(_text_:information in 943) [ClassicSimilarity], result of:
          0.01058886 = score(doc=943,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.116372846 = fieldWeight in 943, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=943)
      0.33333334 = coord(1/3)
    
    Source
    Journal of the Association for Information Science and Technology. 74(2023) no.5, S.570-581
  2. Meng, K.; Ba, Z.; Ma, Y.; Li, G.: ¬A network coupling approach to detecting hierarchical linkages between science and technology (2024) 0.00
    0.0035296201 = product of:
      0.01058886 = sum of:
        0.01058886 = weight(_text_:information in 1205) [ClassicSimilarity], result of:
          0.01058886 = score(doc=1205,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.116372846 = fieldWeight in 1205, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=1205)
      0.33333334 = coord(1/3)
    
    Source
    Journal of the Association for Information Science and Technology. 75(2023) no.2, S.167-187
  3. Kajanan, S.; Bao, Y.; Datta, A.; VanderMeer, D.; Dutta, K.: Efficient automatic search query formulation using phrase-level analysis (2014) 0.00
    0.0033277576 = product of:
      0.009983272 = sum of:
        0.009983272 = weight(_text_:information in 1264) [ClassicSimilarity], result of:
          0.009983272 = score(doc=1264,freq=4.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.10971737 = fieldWeight in 1264, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=1264)
      0.33333334 = coord(1/3)
    
    Abstract
    Over the past decade, the volume of information available digitally over the Internet has grown enormously. Technical developments in the area of search, such as Google's Page Rank algorithm, have proved so good at serving relevant results that Internet search has become integrated into daily human activity. One can endlessly explore topics of interest simply by querying and reading through the resulting links. Yet, although search engines are well known for providing relevant results based on users' queries, users do not always receive the results they are looking for. Google's Director of Research describes clickstream evidence of frustrated users repeatedly reformulating queries and searching through page after page of results. Given the general quality of search engine results, one must consider the possibility that the frustrated user's query is not effective; that is, it does not describe the essence of the user's interest. Indeed, extensive research into human search behavior has found that humans are not very effective at formulating good search queries that describe what they are interested in. Ideally, the user should simply point to a portion of text that sparked the user's interest, and a system should automatically formulate a search query that captures the essence of the text. In this paper, we describe an implemented system that provides this capability. We first describe how our work differs from existing work in automatic query formulation, and propose a new method for improved quantification of the relevance of candidate search terms drawn from input text using phrase-level analysis. We then propose an implementable method designed to provide relevant queries based on a user's text input. We demonstrate the quality of our results and performance of our system through experimental studies. Our results demonstrate that our system produces relevant search terms with roughly two-thirds precision and recall compared to search terms selected by experts, and that typical users find significantly more relevant results (31% more relevant) more quickly (64% faster) using our system than self-formulated search queries. Further, we show that our implementation can scale to request loads of up to 10 requests per second within current online responsiveness expectations (<2-second response times at the highest loads tested).
    Source
    Journal of the Association for Information Science and Technology. 65(2014) no.5, S.1058-1075
  4. Hoenkamp, E.; Bruza, P.D.; Song, D.; Huang, Q.: ¬An effective approach to verbose queries using a limited dependencies language model (2009) 0.00
    0.0033277576 = product of:
      0.009983272 = sum of:
        0.009983272 = weight(_text_:information in 2122) [ClassicSimilarity], result of:
          0.009983272 = score(doc=2122,freq=4.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.10971737 = fieldWeight in 2122, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=2122)
      0.33333334 = coord(1/3)
    
    Series
    Lecture notes in computer science : advances in information retrieval theory; 5766
    Source
    Second International Conference on the Theory of Information Retrieval, ICTIR 2009 Cambridge, UK, September 10-12, 2009 Proceedings. Ed.: L. Azzopardi
  5. Lee, K.H.; Ng, M.K.M.; Lu, Q.: Text segmentation for Chinese spell checking (1999) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 3913) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=3913,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 3913, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3913)
      0.33333334 = coord(1/3)
    
    Source
    Journal of the American Society for Information Science. 50(1999) no.9, S.751-759
  6. Helbig, H.; Gnörlich, C.; Leveling, J.: Natürlichsprachlicher Zugang zu Informationsanbietern im Internet und zu lokalen Datenbanken (2000) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 5558) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=5558,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 5558, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5558)
      0.33333334 = coord(1/3)
    
    Abstract
    Die Schaffung eines natürlichsprachlichen Interfaces (NLI), (das einem Nutzer die Formulierung von Anfragen an Informationsanbieter in seiner Muttersprache erlaubt, stellt eine der interessantesten Herausforderungen im Bereich des Information-Retrieval und der Verarbeitung natürlicher Sprache dar. Dieser Beitrag beschreibt Methoden zur Obersetzung natürlichsprachlicher Anfragen in Ausdrücke formaler Retrievalsprachen sowohl für Informationsressourcen im Internet als auch für lokale Datenbanken. Die vorgestellten Methoden sind Teil das Informationsrecherchesystems LINAS, das an der Fernuniversität Hagen entwickelt wurde, um Nutzern einen natürlichsprachlichen Zugang zu lokalen und zu im Internet verteilten wissenschaftlichen und technischen Informationen anzubieten. Das LINAS-System unterscheidet sich von anderen Systemen und natürlichsprachlichen Interfaces (vgl. OSIRIS) oder die früheren Systeme INTELLECT, Q&A durch die explizite Einbeziehung von Hintergrundwissen und speziellen Dialogmodellen in den Übersetzungsprozeß. Darüber hinaus ist das System auf ein vollständiges Verstehen des natürlichsprachlichen Textes ausgerichtet, während andere Systeme typischerweise nur nach Stichworten oder bestimmten grammatikalischen Mustern in der Eingabe suchen. Ein besonderer Schwerpunkt von LINAS liegt in der Repräsentation und Auswertung der semantischen Relationen zwischen den in der Nutzeranfrage gegebenen Konzepten
  7. Tseng, Y.-H.: Automatic thesaurus generation for Chinese documents (2002) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 5226) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=5226,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 5226, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5226)
      0.33333334 = coord(1/3)
    
    Source
    Journal of the American Society for Information Science and technology. 53(2002) no.13, S.1130-1138
  8. Frederichs, A.: Natürlichsprachige Abfrage und 3-D-Visualisierung von Wissenszusammenhängen (2007) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 566) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=566,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 566, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=566)
      0.33333334 = coord(1/3)
    
    Source
    Wa(h)re Information: 29. Österreichischer Bibliothekartag Bregenz, 19.-23.9.2006. Hrsg.: Harald Weigel
  9. Shaalan, K.; Raza, H.: NERA: Named Entity Recognition for Arabic (2009) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 2953) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=2953,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 2953, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2953)
      0.33333334 = coord(1/3)
    
    Source
    Journal of the American Society for Information Science and Technology. 60(2009) no.8, S.1652-1663
  10. Witschel, H.F.: Text, Wörter, Morpheme : Möglichkeiten einer automatischen Terminologie-Extraktion (2004) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 126) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=126,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 126, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=126)
      0.33333334 = coord(1/3)
    
    Abstract
    Die vorliegende Arbeit beschäftigt sich mit einem Teilgebiet des TextMining, versucht also Information (in diesem Fall Fachterminologie) aus natürlichsprachlichem Text zu extrahieren. Die der Arbeit zugrundeliegende These besagt, daß in vielen Gebieten des Text Mining die Kombination verschiedener Methoden sinnvoll sein kann, um dem Facettenreichtum natürlicher Sprache gerecht zu werden. Die bei der Terminologie-Extraktion angewandten Methoden sind statistischer und linguistischer (bzw. musterbasierter) Natur. Um sie herzuleiten, wurden einige Eigenschaften von Fachtermini herausgearbeitet, die für deren Extraktion relevant sind. So läßt sich z.B. die Tatsache, daß viele Fachbegriffe Nominalphrasen einer bestimmten Form sind, direkt für eine Suche nach gewissen POS-Mustern ausnützen, die Verteilung von Termen in Fachtexten führte zu einem statistischen Ansatz - der Differenzanalyse. Zusammen mit einigen weiteren wurden diese Ansätze in ein Verfahren integriert, welches in der Lage ist, aus dem Feedback eines Anwenders zu lernen und in mehreren Schritten die Suche nach Terminologie zu verfeinern. Dabei wurden mehrere Parameter des Verfahrens veränderlich belassen, d.h. der Anwender kann sie beliebig anpassen. Bei der Untersuchung der Ergebnisse anhand von zwei Fachtexten aus unterschiedlichen Domänen wurde deutlich, daß sich zwar die verschiedenen Verfahren gut ergänzen, daß aber die optimalen Werte der veränderbaren Parameter, ja selbst die Auswahl der angewendeten Verfahren text- und domänenabhängig sind.
  11. Levin, M.; Krawczyk, S.; Bethard, S.; Jurafsky, D.: Citation-based bootstrapping for large-scale author disambiguation (2012) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 246) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=246,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 246, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=246)
      0.33333334 = coord(1/3)
    
    Source
    Journal of the American Society for Information Science and Technology. 63(2012) no.5, S.1030-1047
  12. Fegley, B.D.; Torvik, V.I.: On the role of poetic versus nonpoetic features in "kindred" and diachronic poetry attribution (2012) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 488) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=488,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 488, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=488)
      0.33333334 = coord(1/3)
    
    Source
    Journal of the American Society for Information Science and Technology. 63(2012) no.11, S.2165-2181
  13. Carrillo-de-Albornoz, J.; Plaza, L.: ¬An emotion-based model of negation, intensifiers, and modality for polarity and intensity classification (2013) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 1005) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=1005,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 1005, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1005)
      0.33333334 = coord(1/3)
    
    Source
    Journal of the American Society for Information Science and Technology. 64(2013) no.8, S.1618-1633
  14. Savoy, J.: Text representation strategies : an example with the State of the union addresses (2016) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 3042) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=3042,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 3042, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3042)
      0.33333334 = coord(1/3)
    
    Source
    Journal of the Association for Information Science and Technology. 67(2016) no.8, S.1858-1870
  15. Collovini de Abreu, S.; Vieira, R.: RelP: Portuguese open relation extraction (2017) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 3621) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=3621,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 3621, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3621)
      0.33333334 = coord(1/3)
    
    Abstract
    Natural language texts are valuable data sources in many human activities. NLP techniques are being widely used in order to help find the right information to specific needs. In this paper, we present one such technique: relation extraction from texts. This task aims at identifying and classifying semantic relations that occur between entities in a text. For example, the sentence "Roberto Marinho is the founder of Rede Globo" expresses a relation occurring between "Roberto Marinho" and "Rede Globo." This work presents a system for Portuguese Open Relation Extraction, named RelP, which extracts any relation descriptor that describes an explicit relation between named entities in the organisation domain by applying the Conditional Random Fields. For implementing RelP, we define the representation scheme, features based on previous work, and a reference corpus. RelP achieved state of the art results for open relation extraction; the F-measure rate was around 60% between the named entities person, organisation and place. For better understanding of the output, we present a way for organizing the output from the mining of the extracted relation descriptors. This organization can be useful to classify relation types, to cluster the entities involved in a common relation and to populate datasets.
  16. Kauchak, D.; Leroy, G.; Hogue, A.: Measuring text difficulty using parse-tree frequency (2017) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 3786) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=3786,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 3786, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3786)
      0.33333334 = coord(1/3)
    
    Source
    Journal of the Association for Information Science and Technology. 68(2017) no.9, S.2088-2100
  17. Çelebi, A.; Özgür, A.: Segmenting hashtags and analyzing their grammatical structure (2018) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 4221) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=4221,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 4221, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4221)
      0.33333334 = coord(1/3)
    
    Source
    Journal of the Association for Information Science and Technology. 69(2018) no.5, S.675-686
  18. Muneer, I.; Sharjeel, M.; Iqbal, M.; Adeel Nawab, R.M.; Rayson, P.: CLEU - A Cross-language english-urdu corpus and benchmark for text reuse experiments (2019) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 5299) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=5299,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 5299, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5299)
      0.33333334 = coord(1/3)
    
    Source
    Journal of the Association for Information Science and Technology. 70(2019) no.7, S.729-741
  19. Zhang, Y.; Zhang, C.; Li, J.: Joint modeling of characters, words, and conversation contexts for microblog keyphrase extraction (2020) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 5816) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=5816,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 5816, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5816)
      0.33333334 = coord(1/3)
    
    Source
    Journal of the Association for Information Science and Technology. 71(2020) no.5, S.553-567
  20. Geißler, S.: Natürliche Sprachverarbeitung und Künstliche Intelligenz : ein wachsender Markt mit vielen Chancen. Das Beispiel Kairntech (2020) 0.00
    0.00294135 = product of:
      0.0088240495 = sum of:
        0.0088240495 = weight(_text_:information in 5924) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=5924,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 5924, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5924)
      0.33333334 = coord(1/3)
    
    Source
    Information - Wissenschaft und Praxis. 71(2020) H.2/3, S.95-106

Languages

Types

  • a 394
  • m 35
  • el 26
  • s 19
  • x 10
  • p 3
  • d 2
  • b 1
  • More… Less…

Subjects

Classifications