Search (46 results, page 1 of 3)

  • × theme_ss:"Data Mining"
  1. Hofstede, A.H.M. ter; Proper, H.A.; Van der Weide, T.P.: Exploiting fact verbalisation in conceptual information modelling (1997) 0.03
    0.032544672 = product of:
      0.13017869 = sum of:
        0.13017869 = sum of:
          0.05005701 = weight(_text_:language in 2908) [ClassicSimilarity], result of:
            0.05005701 = score(doc=2908,freq=2.0), product of:
              0.16497234 = queryWeight, product of:
                3.9232929 = idf(docFreq=2376, maxDocs=44218)
                0.042049456 = queryNorm
              0.30342668 = fieldWeight in 2908, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.9232929 = idf(docFreq=2376, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2908)
          0.040241845 = weight(_text_:29 in 2908) [ClassicSimilarity], result of:
            0.040241845 = score(doc=2908,freq=2.0), product of:
              0.14791684 = queryWeight, product of:
                3.5176873 = idf(docFreq=3565, maxDocs=44218)
                0.042049456 = queryNorm
              0.27205724 = fieldWeight in 2908, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5176873 = idf(docFreq=3565, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2908)
          0.039879844 = weight(_text_:22 in 2908) [ClassicSimilarity], result of:
            0.039879844 = score(doc=2908,freq=2.0), product of:
              0.14725003 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.042049456 = queryNorm
              0.2708308 = fieldWeight in 2908, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2908)
      0.25 = coord(1/4)
    
    Abstract
    Focuses on the information modelling side of conceptual modelling. Deals with the exploitation of fact verbalisations after finishing the actual information system. Verbalisations are used as input for the design of the so-called information model. Exploits these verbalisation in 4 directions: considers their use for a conceptual query language, the verbalisation of instances, the description of the contents of a database and for the verbalisation of queries in a computer supported query environment. Provides an example session with an envisioned tool for end user query formulations that exploits the verbalisation
    Date
    5. 4.1996 15:29:15
    Source
    Information systems. 22(1997) nos.5/6, S.349-385
  2. Amir, A.; Feldman, R.; Kashi, R.: ¬A new and versatile method for association generation (1997) 0.02
    0.015261276 = product of:
      0.061045103 = sum of:
        0.061045103 = product of:
          0.09156765 = sum of:
            0.045990683 = weight(_text_:29 in 1270) [ClassicSimilarity], result of:
              0.045990683 = score(doc=1270,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.31092256 = fieldWeight in 1270, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1270)
            0.045576964 = weight(_text_:22 in 1270) [ClassicSimilarity], result of:
              0.045576964 = score(doc=1270,freq=2.0), product of:
                0.14725003 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.042049456 = queryNorm
                0.30952093 = fieldWeight in 1270, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1270)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Date
    5. 4.1996 15:29:15
    Source
    Information systems. 22(1997) nos.5/6, S.333-347
  3. Vaughan, L.; Chen, Y.: Data mining from web search queries : a comparison of Google trends and Baidu index (2015) 0.01
    0.010706769 = product of:
      0.042827077 = sum of:
        0.042827077 = product of:
          0.06424061 = sum of:
            0.03575501 = weight(_text_:language in 1605) [ClassicSimilarity], result of:
              0.03575501 = score(doc=1605,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.21673335 = fieldWeight in 1605, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1605)
            0.028485604 = weight(_text_:22 in 1605) [ClassicSimilarity], result of:
              0.028485604 = score(doc=1605,freq=2.0), product of:
                0.14725003 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.042049456 = queryNorm
                0.19345059 = fieldWeight in 1605, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1605)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Abstract
    Numerous studies have explored the possibility of uncovering information from web search queries but few have examined the factors that affect web query data sources. We conducted a study that investigated this issue by comparing Google Trends and Baidu Index. Data from these two services are based on queries entered by users into Google and Baidu, two of the largest search engines in the world. We first compared the features and functions of the two services based on documents and extensive testing. We then carried out an empirical study that collected query volume data from the two sources. We found that data from both sources could be used to predict the quality of Chinese universities and companies. Despite the differences between the two services in terms of technology, such as differing methods of language processing, the search volume data from the two were highly correlated and combining the two data sources did not improve the predictive power of the data. However, there was a major difference between the two in terms of data availability. Baidu Index was able to provide more search volume data than Google Trends did. Our analysis showed that the disadvantage of Google Trends in this regard was due to Google's smaller user base in China. The implication of this finding goes beyond China. Google's user bases in many countries are smaller than that in China, so the search volume data related to those countries could result in the same issue as that related to China.
    Source
    Journal of the Association for Information Science and Technology. 66(2015) no.1, S.13-22
  4. Budzik, J.; Hammond, K.J.; Birnbaum, L.: Information access in context (2001) 0.01
    0.0067069745 = product of:
      0.026827898 = sum of:
        0.026827898 = product of:
          0.08048369 = sum of:
            0.08048369 = weight(_text_:29 in 3835) [ClassicSimilarity], result of:
              0.08048369 = score(doc=3835,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.5441145 = fieldWeight in 3835, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.109375 = fieldNorm(doc=3835)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    29. 3.2002 17:31:17
  5. Chowdhury, G.G.: Template mining for information extraction from digital documents (1999) 0.01
    0.0066466406 = product of:
      0.026586562 = sum of:
        0.026586562 = product of:
          0.07975969 = sum of:
            0.07975969 = weight(_text_:22 in 4577) [ClassicSimilarity], result of:
              0.07975969 = score(doc=4577,freq=2.0), product of:
                0.14725003 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.042049456 = queryNorm
                0.5416616 = fieldWeight in 4577, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4577)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    2. 4.2000 18:01:22
  6. Witten, I.H.; Frank, E.: Data Mining : Praktische Werkzeuge und Techniken für das maschinelle Lernen (2000) 0.01
    0.0057488354 = product of:
      0.022995342 = sum of:
        0.022995342 = product of:
          0.06898602 = sum of:
            0.06898602 = weight(_text_:29 in 6833) [ClassicSimilarity], result of:
              0.06898602 = score(doc=6833,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.46638384 = fieldWeight in 6833, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6833)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    27. 1.1996 10:29:55
  7. Keim, D.A.: Data Mining mit bloßem Auge (2002) 0.01
    0.0057488354 = product of:
      0.022995342 = sum of:
        0.022995342 = product of:
          0.06898602 = sum of:
            0.06898602 = weight(_text_:29 in 1086) [ClassicSimilarity], result of:
              0.06898602 = score(doc=1086,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.46638384 = fieldWeight in 1086, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=1086)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    31.12.1996 19:29:41
  8. Kruse, R.; Borgelt, C.: Suche im Datendschungel (2002) 0.01
    0.0057488354 = product of:
      0.022995342 = sum of:
        0.022995342 = product of:
          0.06898602 = sum of:
            0.06898602 = weight(_text_:29 in 1087) [ClassicSimilarity], result of:
              0.06898602 = score(doc=1087,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.46638384 = fieldWeight in 1087, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=1087)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    31.12.1996 19:29:41
  9. Wrobel, S.: Lern- und Entdeckungsverfahren (2002) 0.01
    0.0057488354 = product of:
      0.022995342 = sum of:
        0.022995342 = product of:
          0.06898602 = sum of:
            0.06898602 = weight(_text_:29 in 1105) [ClassicSimilarity], result of:
              0.06898602 = score(doc=1105,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.46638384 = fieldWeight in 1105, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=1105)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    31.12.1996 19:29:41
  10. KDD : techniques and applications (1998) 0.01
    0.005697121 = product of:
      0.022788484 = sum of:
        0.022788484 = product of:
          0.06836545 = sum of:
            0.06836545 = weight(_text_:22 in 6783) [ClassicSimilarity], result of:
              0.06836545 = score(doc=6783,freq=2.0), product of:
                0.14725003 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.042049456 = queryNorm
                0.46428138 = fieldWeight in 6783, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6783)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Footnote
    A special issue of selected papers from the Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD'97), held Singapore, 22-23 Feb 1997
  11. Wang, F.L.; Yang, C.C.: Mining Web data for Chinese segmentation (2007) 0.01
    0.0051607913 = product of:
      0.020643165 = sum of:
        0.020643165 = product of:
          0.061929494 = sum of:
            0.061929494 = weight(_text_:language in 604) [ClassicSimilarity], result of:
              0.061929494 = score(doc=604,freq=6.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.3753932 = fieldWeight in 604, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=604)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    Modern information retrieval systems use keywords within documents as indexing terms for search of relevant documents. As Chinese is an ideographic character-based language, the words in the texts are not delimited by white spaces. Indexing of Chinese documents is impossible without a proper segmentation algorithm. Many Chinese segmentation algorithms have been proposed in the past. Traditional segmentation algorithms cannot operate without a large dictionary or a large corpus of training data. Nowadays, the Web has become the largest corpus that is ideal for Chinese segmentation. Although most search engines have problems in segmenting texts into proper words, they maintain huge databases of documents and frequencies of character sequences in the documents. Their databases are important potential resources for segmentation. In this paper, we propose a segmentation algorithm by mining Web data with the help of search engines. On the other hand, the Romanized pinyin of Chinese language indicates boundaries of words in the text. Our algorithm is the first to utilize the Romanized pinyin to segmentation. It is the first unified segmentation algorithm for the Chinese language from different geographical areas, and it is also domain independent because of the nature of the Web. Experiments have been conducted on the datasets of a recent Chinese segmentation competition. The results show that our algorithm outperforms the traditional algorithms in terms of precision and recall. Moreover, our algorithm can effectively deal with the problems of segmentation ambiguity, new word (unknown word) detection, and stop words.
  12. Teich, E.; Degaetano-Ortlieb, S.; Fankhauser, P.; Kermes, H.; Lapshinova-Koltunski, E.: ¬The linguistic construal of disciplinarity : a data-mining approach using register features (2016) 0.01
    0.005056522 = product of:
      0.020226087 = sum of:
        0.020226087 = product of:
          0.06067826 = sum of:
            0.06067826 = weight(_text_:language in 3015) [ClassicSimilarity], result of:
              0.06067826 = score(doc=3015,freq=4.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.3678087 = fieldWeight in 3015, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3015)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    We analyze the linguistic evolution of selected scientific disciplines over a 30-year time span (1970s to 2000s). Our focus is on four highly specialized disciplines at the boundaries of computer science that emerged during that time: computational linguistics, bioinformatics, digital construction, and microelectronics. Our analysis is driven by the question whether these disciplines develop a distinctive language use-both individually and collectively-over the given time period. The data set is the English Scientific Text Corpus (scitex), which includes texts from the 1970s/1980s and early 2000s. Our theoretical basis is register theory. In terms of methods, we combine corpus-based methods of feature extraction (various aggregated features [part-of-speech based], n-grams, lexico-grammatical patterns) and automatic text classification. The results of our research are directly relevant to the study of linguistic variation and languages for specific purposes (LSP) and have implications for various natural language processing (NLP) tasks, for example, authorship attribution, text mining, or training NLP tools.
  13. Borgelt, C.; Kruse, R.: Unsicheres Wissen nutzen (2002) 0.00
    0.0047906963 = product of:
      0.019162785 = sum of:
        0.019162785 = product of:
          0.057488356 = sum of:
            0.057488356 = weight(_text_:29 in 1104) [ClassicSimilarity], result of:
              0.057488356 = score(doc=1104,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.38865322 = fieldWeight in 1104, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1104)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    31.12.1996 19:29:41
  14. Tonkin, E.L.; Tourte, G.J.L.: Working with text. tools, techniques and approaches for text mining (2016) 0.00
    0.004213768 = product of:
      0.016855072 = sum of:
        0.016855072 = product of:
          0.050565217 = sum of:
            0.050565217 = weight(_text_:language in 4019) [ClassicSimilarity], result of:
              0.050565217 = score(doc=4019,freq=4.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.30650726 = fieldWeight in 4019, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4019)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    What is text mining, and how can it be used? What relevance do these methods have to everyday work in information science and the digital humanities? How does one develop competences in text mining? Working with Text provides a series of cross-disciplinary perspectives on text mining and its applications. As text mining raises legal and ethical issues, the legal background of text mining and the responsibilities of the engineer are discussed in this book. Chapters provide an introduction to the use of the popular GATE text mining package with data drawn from social media, the use of text mining to support semantic search, the development of an authority system to support content tagging, and recent techniques in automatic language evaluation. Focused studies describe text mining on historical texts, automated indexing using constrained vocabularies, and the use of natural language processing to explore the climate science literature. Interviews are included that offer a glimpse into the real-life experience of working within commercial and academic text mining.
  15. Search tools (1997) 0.00
    0.0041714176 = product of:
      0.01668567 = sum of:
        0.01668567 = product of:
          0.05005701 = sum of:
            0.05005701 = weight(_text_:language in 3834) [ClassicSimilarity], result of:
              0.05005701 = score(doc=3834,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.30342668 = fieldWeight in 3834, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3834)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    Offers brief accounts of Internet search tools. Covers the Lycos revamp; the new navigation service produced jointly by Excite and Netscape, delivering a language specific, locally relevant Web guide for Japan, Germany, France, the UK and Australia; InfoWatcher, a combination offline browser, search engine and push product from Carvelle Inc., USA; Alexa by Alexa Internet and WBI from IBM which are free and provide users with information on how others have used the Web sites which they are visiting; and Concept Explorer from Knowledge Discovery Systems, Inc., California which performs data mining from the Web, Usenet groups, MEDLINE and the US Patent and Trademark Office patent abstracts
  16. Lam, W.; Yang, C.C.; Menczer, F.: Introduction to the special topic section on mining Web resources for enhancing information retrieval (2007) 0.00
    0.0041714176 = product of:
      0.01668567 = sum of:
        0.01668567 = product of:
          0.05005701 = sum of:
            0.05005701 = weight(_text_:language in 600) [ClassicSimilarity], result of:
              0.05005701 = score(doc=600,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.30342668 = fieldWeight in 600, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=600)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Abstract
    The amount of information on the Web has been expanding at an enormous pace. There are a variety of Web documents in different genres, such as news, reports, reviews. Traditionally, the information displayed on Web sites has been static. Recently, there are many Web sites offering content that is dynamically generated and frequently updated. It is also common for Web sites to contain information in different languages since many countries adopt more than one language. Moreover, content may exist in multimedia formats including text, images, video, and audio.
  17. Cardie, C.: Empirical methods in information extraction (1997) 0.00
    0.003832557 = product of:
      0.015330228 = sum of:
        0.015330228 = product of:
          0.045990683 = sum of:
            0.045990683 = weight(_text_:29 in 3246) [ClassicSimilarity], result of:
              0.045990683 = score(doc=3246,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.31092256 = fieldWeight in 3246, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3246)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    6. 3.1999 13:50:29
  18. Tiefschürfen in Datenbanken (2002) 0.00
    0.003832557 = product of:
      0.015330228 = sum of:
        0.015330228 = product of:
          0.045990683 = sum of:
            0.045990683 = weight(_text_:29 in 996) [ClassicSimilarity], result of:
              0.045990683 = score(doc=996,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.31092256 = fieldWeight in 996, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=996)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    31.12.1996 19:29:41
  19. Bath, P.A.: Data mining in health and medical information (2003) 0.00
    0.003832557 = product of:
      0.015330228 = sum of:
        0.015330228 = product of:
          0.045990683 = sum of:
            0.045990683 = weight(_text_:29 in 4263) [ClassicSimilarity], result of:
              0.045990683 = score(doc=4263,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.31092256 = fieldWeight in 4263, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4263)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    23.10.2005 18:29:03
  20. Matson, L.D.; Bonski, D.J.: Do digital libraries need librarians? (1997) 0.00
    0.0037980804 = product of:
      0.0151923215 = sum of:
        0.0151923215 = product of:
          0.045576964 = sum of:
            0.045576964 = weight(_text_:22 in 1737) [ClassicSimilarity], result of:
              0.045576964 = score(doc=1737,freq=2.0), product of:
                0.14725003 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.042049456 = queryNorm
                0.30952093 = fieldWeight in 1737, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1737)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Date
    22.11.1998 18:57:22

Years

Languages

  • e 31
  • d 15

Types