Search (350 results, page 18 of 18)

  • × type_ss:"el"
  1. Bertolucci, K.: Happiness is taxonomy : four structures for Snoopy - libraries' method of categorizing and classification (2003) 0.00
    0.0048741973 = product of:
      0.009748395 = sum of:
        0.009748395 = product of:
          0.01949679 = sum of:
            0.01949679 = weight(_text_:p in 1212) [ClassicSimilarity], result of:
              0.01949679 = score(doc=1212,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.11917553 = fieldWeight in 1212, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1212)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Content
    Vgl.: http://findarticles.com/p/articles/mi_m0FWE/is_3_7/ai_99011617.
  2. Beuth, P.: Voyeure gesucht : Böse Nachbarn (2008) 0.00
    0.0048741973 = product of:
      0.009748395 = sum of:
        0.009748395 = product of:
          0.01949679 = sum of:
            0.01949679 = weight(_text_:p in 2226) [ClassicSimilarity], result of:
              0.01949679 = score(doc=2226,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.11917553 = fieldWeight in 2226, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=2226)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  3. Adler, R.; Ewing, J.; Taylor, P.: Citation statistics : A report from the International Mathematical Union (IMU) in cooperation with the International Council of Industrial and Applied Mathematics (ICIAM) and the Institute of Mathematical Statistics (IMS) (2008) 0.00
    0.0048741973 = product of:
      0.009748395 = sum of:
        0.009748395 = product of:
          0.01949679 = sum of:
            0.01949679 = weight(_text_:p in 2417) [ClassicSimilarity], result of:
              0.01949679 = score(doc=2417,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.11917553 = fieldWeight in 2417, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=2417)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  4. Foerster, H. von; Müller, A.; Müller, K.H.: Rück- und Vorschauen : Heinz von Foerster im Gespräch mit Albert Müller und Karl H. Müller (2001) 0.00
    0.0046235067 = product of:
      0.009247013 = sum of:
        0.009247013 = product of:
          0.018494027 = sum of:
            0.018494027 = weight(_text_:22 in 5988) [ClassicSimilarity], result of:
              0.018494027 = score(doc=5988,freq=2.0), product of:
                0.15933464 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045500398 = queryNorm
                0.116070345 = fieldWeight in 5988, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=5988)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    10. 9.2006 17:22:54
  5. Lavoie, B.; Connaway, L.S.; Dempsey, L.: Anatomy of aggregate collections : the example of Google print for libraries (2005) 0.00
    0.0046235067 = product of:
      0.009247013 = sum of:
        0.009247013 = product of:
          0.018494027 = sum of:
            0.018494027 = weight(_text_:22 in 1184) [ClassicSimilarity], result of:
              0.018494027 = score(doc=1184,freq=2.0), product of:
                0.15933464 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045500398 = queryNorm
                0.116070345 = fieldWeight in 1184, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1184)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    26.12.2011 14:08:22
  6. Graphic details : a scientific study of the importance of diagrams to science (2016) 0.00
    0.0046235067 = product of:
      0.009247013 = sum of:
        0.009247013 = product of:
          0.018494027 = sum of:
            0.018494027 = weight(_text_:22 in 3035) [ClassicSimilarity], result of:
              0.018494027 = score(doc=3035,freq=2.0), product of:
                0.15933464 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045500398 = queryNorm
                0.116070345 = fieldWeight in 3035, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=3035)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Content
    As the team describe in a paper posted (http://arxiv.org/abs/1605.04951) on arXiv, they found that figures did indeed matter-but not all in the same way. An average paper in PubMed Central has about one diagram for every three pages and gets 1.67 citations. Papers with more diagrams per page and, to a lesser extent, plots per page tended to be more influential (on average, a paper accrued two more citations for every extra diagram per page, and one more for every extra plot per page). By contrast, including photographs and equations seemed to decrease the chances of a paper being cited by others. That agrees with a study from 2012, whose authors counted (by hand) the number of mathematical expressions in over 600 biology papers and found that each additional equation per page reduced the number of citations a paper received by 22%. This does not mean that researchers should rush to include more diagrams in their next paper. Dr Howe has not shown what is behind the effect, which may merely be one of correlation, rather than causation. It could, for example, be that papers with lots of diagrams tend to be those that illustrate new concepts, and thus start a whole new field of inquiry. Such papers will certainly be cited a lot. On the other hand, the presence of equations really might reduce citations. Biologists (as are most of those who write and read the papers in PubMed Central) are notoriously mathsaverse. If that is the case, looking in a physics archive would probably produce a different result.
  7. DeSilva, J.M.; Traniello, J.F.A.; Claxton, A.G.; Fannin, L.D.: When and why did human brains decrease in size? : a new change-point analysis and insights from brain evolution in ants (2021) 0.00
    0.0046235067 = product of:
      0.009247013 = sum of:
        0.009247013 = product of:
          0.018494027 = sum of:
            0.018494027 = weight(_text_:22 in 405) [ClassicSimilarity], result of:
              0.018494027 = score(doc=405,freq=2.0), product of:
                0.15933464 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045500398 = queryNorm
                0.116070345 = fieldWeight in 405, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=405)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Frontiers in ecology and evolution, 22 October 2021 [https://www.frontiersin.org/articles/10.3389/fevo.2021.742639/full]
  8. Beuth, P.: ¬Das Netz der Welt : Lobos Webciety (2009) 0.00
    0.004061831 = product of:
      0.008123662 = sum of:
        0.008123662 = product of:
          0.016247325 = sum of:
            0.016247325 = weight(_text_:p in 2136) [ClassicSimilarity], result of:
              0.016247325 = score(doc=2136,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.099312946 = fieldWeight in 2136, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=2136)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  9. Crane, G.; Jones, A.: Text, information, knowledge and the evolving record of humanity (2006) 0.00
    0.004061831 = product of:
      0.008123662 = sum of:
        0.008123662 = product of:
          0.016247325 = sum of:
            0.016247325 = weight(_text_:p in 1182) [ClassicSimilarity], result of:
              0.016247325 = score(doc=1182,freq=2.0), product of:
                0.16359726 = queryWeight, product of:
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.045500398 = queryNorm
                0.099312946 = fieldWeight in 1182, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5955126 = idf(docFreq=3298, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=1182)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Consider a sentence such as "the current price of tea in China is 35 cents per pound." In a library with millions of books we might find many statements of the above form that we could capture today with relatively simple rules: rather than pursuing every variation of a statement, programs can wait, like predators at a water hole, for their informational prey to reappear in a standard linguistic pattern. We can make inferences from sentences such as "NAME1 born at NAME2 in DATE" that NAME more likely than not represents a person and NAME a place and then convert the statement into a proposition about a person born at a given place and time. The changing price of tea in China, pedestrian birth and death dates, or other basic statements may not be truth and beauty in the Phaedrus, but a digital library that could plot the prices of various commodities in different markets over time, plot the various lifetimes of individuals, or extract and classify many events would be very useful. Services such as the Syllabus Finder1 and H-Bot2 (which Dan Cohen describes elsewhere in this issue of D-Lib) represent examples of information extraction already in use. H-Bot, in particular, builds on our evolving ability to extract information from very large corpora such as the billions of web pages available through the Google API. Aside from identifying higher order statements, however, users also want to search and browse named entities: they want to read about "C. P. E. Bach" rather than his father "Johann Sebastian" or about "Cambridge, Maryland", without hearing about "Cambridge, Massachusetts", Cambridge in the UK or any of the other Cambridges scattered around the world. Named entity identification is a well-established area with an ongoing literature. The Natural Language Processing Research Group at the University of Sheffield has developed its open source Generalized Architecture for Text Engineering (GATE) for years, while IBM's Unstructured Information Analysis and Search (UIMA) is "available as open source software to provide a common foundation for industry and academia." Powerful tools are thus freely available and more demanding users can draw upon published literature to develop their own systems. Major search engines such as Google and Yahoo also integrate increasingly sophisticated tools to categorize and identify places. The software resources are rich and expanding. The reference works on which these systems depend, however, are ill-suited for historical analysis. First, simple gazetteers and similar authority lists quickly grow too big for useful information extraction. They provide us with potential entities against which to match textual references, but existing electronic reference works assume that human readers can use their knowledge of geography and of the immediate context to pick the right Boston from the Bostons in the Getty Thesaurus of Geographic Names (TGN), but, with the crucial exception of geographic location, the TGN records do not provide any machine readable clues: we cannot tell which Bostons are large or small. If we are analyzing a document published in 1818, we cannot filter out those places that did not yet exist or that had different names: "Jefferson Davis" is not the name of a parish in Louisiana (tgn,2000880) or a county in Mississippi (tgn,2001118) until after the Civil War.
  10. Laaff, M.: Googles genialer Urahn (2011) 0.00
    0.0038529225 = product of:
      0.007705845 = sum of:
        0.007705845 = product of:
          0.01541169 = sum of:
            0.01541169 = weight(_text_:22 in 4610) [ClassicSimilarity], result of:
              0.01541169 = score(doc=4610,freq=2.0), product of:
                0.15933464 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045500398 = queryNorm
                0.09672529 = fieldWeight in 4610, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=4610)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    24.10.2008 14:19:22

Years

Languages

  • e 202
  • d 138
  • el 2
  • a 1
  • f 1
  • i 1
  • nl 1
  • More… Less…

Types

  • a 149
  • p 34
  • i 11
  • r 8
  • m 6
  • x 5
  • s 4
  • n 3
  • b 2
  • More… Less…