Search (26 results, page 1 of 2)

  • × theme_ss:"Computerlinguistik"
  • × year_i:[2010 TO 2020}
  1. Engerer, V.: Indexierungstheorie für Linguisten : zu einigen natürlichsprachlichen Zügen in künstlichen Indexsprachen (2014) 0.09
    0.09407521 = product of:
      0.18815042 = sum of:
        0.12738486 = weight(_text_:et in 3339) [ClassicSimilarity], result of:
          0.12738486 = score(doc=3339,freq=2.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.6220778 = fieldWeight in 3339, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.09375 = fieldNorm(doc=3339)
        0.06076556 = product of:
          0.12153112 = sum of:
            0.12153112 = weight(_text_:al in 3339) [ClassicSimilarity], result of:
              0.12153112 = score(doc=3339,freq=2.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.60761654 = fieldWeight in 3339, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3339)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Source
    Dialekte, Konzepte, Kontakte. Ergebnisse des Arbeitstreffens der Gesellschaft für Sprache und Sprachen, GeSuS e.V., 31. Mai - 1. Juni 2013 in Freiburg/Breisgau. Hrsg.: V. Schönenberger et al
  2. Zadeh, B.Q.; Handschuh, S.: ¬The ACL RD-TEC : a dataset for benchmarking terminology extraction and classification in computational linguistics (2014) 0.05
    0.047037605 = product of:
      0.09407521 = sum of:
        0.06369243 = weight(_text_:et in 2803) [ClassicSimilarity], result of:
          0.06369243 = score(doc=2803,freq=2.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.3110389 = fieldWeight in 2803, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.046875 = fieldNorm(doc=2803)
        0.03038278 = product of:
          0.06076556 = sum of:
            0.06076556 = weight(_text_:al in 2803) [ClassicSimilarity], result of:
              0.06076556 = score(doc=2803,freq=2.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.30380827 = fieldWeight in 2803, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2803)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Source
    Proceedings of the 4th International Workshop on Computational Terminology, Dublin, Ireland, August 23 2014. COLING 2014. Eds.: Patrick Drouin et al., Dublin, Ireland, 2014-08-23 [https://www.deri.ie/sites/default/files/publications/the-acl-rd-tec.pdf]
  3. Korman, D.Z.; Mack, E.; Jett, J.; Renear, A.H.: Defining textual entailment (2018) 0.05
    0.047037605 = product of:
      0.09407521 = sum of:
        0.06369243 = weight(_text_:et in 4284) [ClassicSimilarity], result of:
          0.06369243 = score(doc=4284,freq=2.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.3110389 = fieldWeight in 4284, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.046875 = fieldNorm(doc=4284)
        0.03038278 = product of:
          0.06076556 = sum of:
            0.06076556 = weight(_text_:al in 4284) [ClassicSimilarity], result of:
              0.06076556 = score(doc=4284,freq=2.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.30380827 = fieldWeight in 4284, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4284)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Textual entailment is a relationship that obtains between fragments of text when one fragment in some sense implies the other fragment. The automation of textual entailment recognition supports a wide variety of text-based tasks, including information retrieval, information extraction, question answering, text summarization, and machine translation. Much ingenuity has been devoted to developing algorithms for identifying textual entailments, but relatively little to saying what textual entailment actually is. This article is a review of the logical and philosophical issues involved in providing an adequate definition of textual entailment. We show that many natural definitions of textual entailment are refuted by counterexamples, including the most widely cited definition of Dagan et al. We then articulate and defend the following revised definition: T textually entails H?=?df typically, a human reading T would be justified in inferring the proposition expressed by H from the proposition expressed by T. We also show that textual entailment is context-sensitive, nontransitive, and nonmonotonic.
  4. Rozinajová, V.; Macko, P.: Using natural language to search linked data (2017) 0.04
    0.039198004 = product of:
      0.07839601 = sum of:
        0.053077027 = weight(_text_:et in 3488) [ClassicSimilarity], result of:
          0.053077027 = score(doc=3488,freq=2.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.2591991 = fieldWeight in 3488, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3488)
        0.025318984 = product of:
          0.050637968 = sum of:
            0.050637968 = weight(_text_:al in 3488) [ClassicSimilarity], result of:
              0.050637968 = score(doc=3488,freq=2.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.25317356 = fieldWeight in 3488, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3488)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Source
    Semantic keyword-based search on structured data sources: COST Action IC1302. Second International KEYSTONE Conference, IKC 2016, Cluj-Napoca, Romania, September 8-9, 2016, Revised Selected Papers. Eds.: A. Calì, A. et al
  5. Nagy T., I.: Detecting multiword expressions and named entities in natural language texts (2014) 0.04
    0.038804047 = product of:
      0.07760809 = sum of:
        0.052543577 = weight(_text_:et in 1536) [ClassicSimilarity], result of:
          0.052543577 = score(doc=1536,freq=4.0), product of:
            0.20477319 = queryWeight, product of:
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.043643 = queryNorm
            0.25659403 = fieldWeight in 1536, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.692005 = idf(docFreq=1101, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1536)
        0.025064515 = product of:
          0.05012903 = sum of:
            0.05012903 = weight(_text_:al in 1536) [ClassicSimilarity], result of:
              0.05012903 = score(doc=1536,freq=4.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.25062904 = fieldWeight in 1536, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1536)
          0.5 = coord(1/2)
      0.5 = coord(2/4)
    
    Abstract
    Multiword expressions (MWEs) are lexical items that can be decomposed into single words and display lexical, syntactic, semantic, pragmatic and/or statistical idiosyncrasy (Sag et al., 2002; Kim, 2008; Calzolari et al., 2002). The proper treatment of multiword expressions such as rock 'n' roll and make a decision is essential for many natural language processing (NLP) applications like information extraction and retrieval, terminology extraction and machine translation, and it is important to identify multiword expressions in context. For example, in machine translation we must know that MWEs form one semantic unit, hence their parts should not be translated separately. For this, multiword expressions should be identified first in the text to be translated. The chief aim of this thesis is to develop machine learning-based approaches for the automatic detection of different types of multiword expressions in English and Hungarian natural language texts. In our investigations, we pay attention to the characteristics of different types of multiword expressions such as nominal compounds, multiword named entities and light verb constructions, and we apply novel methods to identify MWEs in raw texts. In the thesis it will be demonstrated that nominal compounds and multiword amed entities may require a similar approach for their automatic detection as they behave in the same way from a linguistic point of view. Furthermore, it will be shown that the automatic detection of light verb constructions can be carried out using two effective machine learning-based approaches.
  6. Kocijan, K.: Visualizing natural language resources (2015) 0.02
    0.020284442 = product of:
      0.08113777 = sum of:
        0.08113777 = weight(_text_:c in 2995) [ClassicSimilarity], result of:
          0.08113777 = score(doc=2995,freq=4.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.5389696 = fieldWeight in 2995, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.078125 = fieldNorm(doc=2995)
      0.25 = coord(1/4)
    
    Source
    Re:inventing information science in the networked society: Proceedings of the 14th International Symposium on Information Science, Zadar/Croatia, 19th-21st May 2015. Eds.: F. Pehar, C. Schloegl u. C. Wolff
  7. Al-Shawakfa, E.; Al-Badarneh, A.; Shatnawi, S.; Al-Rabab'ah, K.; Bani-Ismail, B.: ¬A comparison study of some Arabic root finding algorithms (2010) 0.01
    0.013156131 = product of:
      0.052624524 = sum of:
        0.052624524 = product of:
          0.10524905 = sum of:
            0.10524905 = weight(_text_:al in 3457) [ClassicSimilarity], result of:
              0.10524905 = score(doc=3457,freq=6.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.5262114 = fieldWeight in 3457, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3457)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
  8. Lu, C.; Bu, Y.; Wang, J.; Ding, Y.; Torvik, V.; Schnaars, M.; Zhang, C.: Examining scientific writing styles from the perspective of linguistic complexity : a cross-level moderation model (2019) 0.01
    0.012170665 = product of:
      0.04868266 = sum of:
        0.04868266 = weight(_text_:c in 5219) [ClassicSimilarity], result of:
          0.04868266 = score(doc=5219,freq=4.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.32338172 = fieldWeight in 5219, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.046875 = fieldNorm(doc=5219)
      0.25 = coord(1/4)
    
  9. Becks, D.; Schulz, J.M.: Domänenübergreifende Phrasenextraktion mithilfe einer lexikonunabhängigen Analysekomponente (2010) 0.01
    0.011474612 = product of:
      0.04589845 = sum of:
        0.04589845 = weight(_text_:c in 4661) [ClassicSimilarity], result of:
          0.04589845 = score(doc=4661,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.3048872 = fieldWeight in 4661, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.0625 = fieldNorm(doc=4661)
      0.25 = coord(1/4)
    
    Source
    Information und Wissen: global, sozial und frei? Proceedings des 12. Internationalen Symposiums für Informationswissenschaft (ISI 2011) ; Hildesheim, 9. - 11. März 2011. Hrsg.: J. Griesbaum, T. Mandl u. C. Womser-Hacker
  10. Hmeidi, I.I.; Al-Shalabi, R.F.; Al-Taani, A.T.; Najadat, H.; Al-Hazaimeh, S.A.: ¬A novel approach to the extraction of roots from Arabic words using bigrams (2010) 0.01
    0.010963443 = product of:
      0.04385377 = sum of:
        0.04385377 = product of:
          0.08770754 = sum of:
            0.08770754 = weight(_text_:al in 3426) [ClassicSimilarity], result of:
              0.08770754 = score(doc=3426,freq=6.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.43850952 = fieldWeight in 3426, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3426)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
  11. AL-Smadi, M.; Jaradat, Z.; AL-Ayyoub, M.; Jararweh, Y.: Paraphrase identification and semantic text similarity analysis in Arabic news tweets using lexical, syntactic, and semantic features (2017) 0.01
    0.010741935 = product of:
      0.04296774 = sum of:
        0.04296774 = product of:
          0.08593548 = sum of:
            0.08593548 = weight(_text_:al in 5095) [ClassicSimilarity], result of:
              0.08593548 = score(doc=5095,freq=4.0), product of:
                0.20001286 = queryWeight, product of:
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.043643 = queryNorm
                0.42964977 = fieldWeight in 5095, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.582931 = idf(docFreq=1228, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5095)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
  12. Vasalou, A.; Gill, A.J.; Mazanderani, F.; Papoutsi, C.; Joinson, A.: Privacy dictionary : a new resource for the automated content analysis of privacy (2011) 0.01
    0.00860596 = product of:
      0.03442384 = sum of:
        0.03442384 = weight(_text_:c in 4915) [ClassicSimilarity], result of:
          0.03442384 = score(doc=4915,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.22866541 = fieldWeight in 4915, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.046875 = fieldNorm(doc=4915)
      0.25 = coord(1/4)
    
  13. Ramisch, C.; Villavicencio, A.; Kordoni, V.: Introduction to the special issue on multiword expressions : from theory to practice and use (2013) 0.01
    0.00860596 = product of:
      0.03442384 = sum of:
        0.03442384 = weight(_text_:c in 1124) [ClassicSimilarity], result of:
          0.03442384 = score(doc=1124,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.22866541 = fieldWeight in 1124, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.046875 = fieldNorm(doc=1124)
      0.25 = coord(1/4)
    
  14. Rosemblat, G.; Resnick, M.P.; Auston, I.; Shin, D.; Sneiderman, C.; Fizsman, M.; Rindflesch, T.C.: Extending SemRep to the public health domain (2013) 0.01
    0.00860596 = product of:
      0.03442384 = sum of:
        0.03442384 = weight(_text_:c in 2096) [ClassicSimilarity], result of:
          0.03442384 = score(doc=2096,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.22866541 = fieldWeight in 2096, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.046875 = fieldNorm(doc=2096)
      0.25 = coord(1/4)
    
  15. Anguiano Peña, G.; Naumis Peña, C.: Method for selecting specialized terms from a general language corpus (2015) 0.01
    0.00860596 = product of:
      0.03442384 = sum of:
        0.03442384 = weight(_text_:c in 2196) [ClassicSimilarity], result of:
          0.03442384 = score(doc=2196,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.22866541 = fieldWeight in 2196, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.046875 = fieldNorm(doc=2196)
      0.25 = coord(1/4)
    
  16. Malo, P.; Sinha, A.; Korhonen, P.; Wallenius, J.; Takala, P.: Good debt or bad debt : detecting semantic orientations in economic texts (2014) 0.01
    0.0071716327 = product of:
      0.02868653 = sum of:
        0.02868653 = weight(_text_:c in 1226) [ClassicSimilarity], result of:
          0.02868653 = score(doc=1226,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.1905545 = fieldWeight in 1226, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1226)
      0.25 = coord(1/4)
    
    Abstract
    The use of robo-readers to analyze news texts is an emerging technology trend in computational finance. Recent research has developed sophisticated financial polarity lexicons for investigating how financial sentiments relate to future company performance. However, based on experience from fields that commonly analyze sentiment, it is well known that the overall semantic orientation of a sentence may differ from that of individual words. This article investigates how semantic orientations can be better detected in financial and economic news by accommodating the overall phrase-structure information and domain-specific use of language. Our three main contributions are the following: (a) a human-annotated finance phrase bank that can be used for training and evaluating alternative models; (b) a technique to enhance financial lexicons with attributes that help to identify expected direction of events that affect sentiment; and (c) a linearized phrase-structure model for detecting contextual semantic orientations in economic texts. The relevance of the newly added lexicon features and the benefit of using the proposed learning algorithm are demonstrated in a comparative study against general sentiment models as well as the popular word frequency models used in recent financial studies. The proposed framework is parsimonious and avoids the explosion in feature space caused by the use of conventional n-gram features.
  17. Lian, T.; Yu, C.; Wang, W.; Yuan, Q.; Hou, Z.: Doctoral dissertations on tourism in China : a co-word analysis (2016) 0.01
    0.0071716327 = product of:
      0.02868653 = sum of:
        0.02868653 = weight(_text_:c in 3178) [ClassicSimilarity], result of:
          0.02868653 = score(doc=3178,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.1905545 = fieldWeight in 3178, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3178)
      0.25 = coord(1/4)
    
  18. Doval, Y.; Gómez-Rodríguez, C.: Comparing neural- and N-gram-based language models for word segmentation (2019) 0.01
    0.0071716327 = product of:
      0.02868653 = sum of:
        0.02868653 = weight(_text_:c in 4675) [ClassicSimilarity], result of:
          0.02868653 = score(doc=4675,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.1905545 = fieldWeight in 4675, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4675)
      0.25 = coord(1/4)
    
  19. Lezius, W.: Morphy - Morphologie und Tagging für das Deutsche (2013) 0.01
    0.0059130243 = product of:
      0.023652097 = sum of:
        0.023652097 = product of:
          0.047304194 = sum of:
            0.047304194 = weight(_text_:22 in 1490) [ClassicSimilarity], result of:
              0.047304194 = score(doc=1490,freq=2.0), product of:
                0.15283036 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.043643 = queryNorm
                0.30952093 = fieldWeight in 1490, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1490)
          0.5 = coord(1/2)
      0.25 = coord(1/4)
    
    Date
    22. 3.2015 9:30:24
  20. Vlachidis, A.; Binding, C.; Tudhope, D.; May, K.: Excavating grey literature : a case study on the rich indexing of archaeological documents via natural language-processing techniques and knowledge-based resources (2010) 0.01
    0.005737306 = product of:
      0.022949224 = sum of:
        0.022949224 = weight(_text_:c in 3948) [ClassicSimilarity], result of:
          0.022949224 = score(doc=3948,freq=2.0), product of:
            0.1505424 = queryWeight, product of:
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.043643 = queryNorm
            0.1524436 = fieldWeight in 3948, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.4494052 = idf(docFreq=3817, maxDocs=44218)
              0.03125 = fieldNorm(doc=3948)
      0.25 = coord(1/4)