Search (69 results, page 2 of 4)

  • × theme_ss:"Internet"
  • × theme_ss:"Suchmaschinen"
  • × type_ss:"a"
  1. Jacobs, N.A.; Buden, I.D.; Discombe, R.J.: Researchers on the pier : academics' use of a local Internet gateway (1996) 0.01
    0.0072560436 = product of:
      0.01814011 = sum of:
        0.012614433 = weight(_text_:a in 2605) [ClassicSimilarity], result of:
          0.012614433 = score(doc=2605,freq=14.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.23593865 = fieldWeight in 2605, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2605)
        0.005525676 = product of:
          0.011051352 = sum of:
            0.011051352 = weight(_text_:information in 2605) [ClassicSimilarity], result of:
              0.011051352 = score(doc=2605,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.13576832 = fieldWeight in 2605, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2605)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Describes the establishment of an Internet gateway local to the University of sussex, UK, firstly as a gopher and more recently as a WWW server. Describes a project currently uderway to assess the practical value of Internet searches to research and the consequential training and support requirements. Determines the value of a local gateway to academic researchers at Sussex and the roles specific to a local gateway that such comprehensive resources as Yahoo and BUBL cannot fulfil. The project is using an interview-based methodology to gain an understanding of researchers' views of Internet resource provision, and these interviews are being analyzed with the help of the NUD*IST qualitative data analysis software package
    Source
    New review of information networking. 1996, no.2, S.149-167
    Type
    a
  2. Rowlands, I.; Nicholas, D.; Williams, P.; Huntington, P.; Fieldhouse, M.; Gunter, B.; Withey, R.; Jamali, H.R.; Dobrowolski, T.; Tenopir, C.: ¬The Google generation : the information behaviour of the researcher of the future (2008) 0.01
    0.007134348 = product of:
      0.01783587 = sum of:
        0.009010308 = weight(_text_:a in 2017) [ClassicSimilarity], result of:
          0.009010308 = score(doc=2017,freq=14.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.1685276 = fieldWeight in 2017, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2017)
        0.008825562 = product of:
          0.017651124 = sum of:
            0.017651124 = weight(_text_:information in 2017) [ClassicSimilarity], result of:
              0.017651124 = score(doc=2017,freq=10.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.21684799 = fieldWeight in 2017, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2017)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Purpose - This article is an edited version of a report commissioned by the British Library and JISC to identify how the specialist researchers of the future (those born after 1993) are likely to access and interact with digital resources in five to ten years' time. The purpose is to investigate the impact of digital transition on the information behaviour of the Google Generation and to guide library and information services to anticipate and react to any new or emerging behaviours in the most effective way. Design/methodology/approach - The study was virtually longitudinal and is based on a number of extensive reviews of related literature, survey data mining and a deep log analysis of a British Library and a JISC web site intended for younger people. Findings - The study shows that much of the impact of ICTs on the young has been overestimated. The study claims that although young people demonstrate an apparent ease and familiarity with computers, they rely heavily on search engines, view rather than read and do not possess the critical and analytical skills to assess the information that they find on the web. Originality/value - The paper reports on a study that overturns the common assumption that the "Google generation" is the most web-literate.
    Footnote
    Vgl. auch: Rowlands, I.: Google generation: issues in information literacy. In: http://www.lucis.me.uk/retrieval%20issues.pdf.
    Type
    a
  3. Hilberer, T.: Über die Zugänglichkeit der Informationen im Internet : Die Rolle der Bibliotheken (1999) 0.01
    0.007058388 = product of:
      0.01764597 = sum of:
        0.008173384 = weight(_text_:a in 4101) [ClassicSimilarity], result of:
          0.008173384 = score(doc=4101,freq=2.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.15287387 = fieldWeight in 4101, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.09375 = fieldNorm(doc=4101)
        0.009472587 = product of:
          0.018945174 = sum of:
            0.018945174 = weight(_text_:information in 4101) [ClassicSimilarity], result of:
              0.018945174 = score(doc=4101,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.23274569 = fieldWeight in 4101, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4101)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Footnote
    Bezugnahme auf den Artikel: Lawrence, S. u. C.L. Giles: Accessibility of information on the web. In: Nature. No.400 vom 8.7.1999, S.107-109.
    Type
    a
  4. Butler, D.: Souped-up search engines (2000) 0.01
    0.0070104985 = product of:
      0.017526247 = sum of:
        0.009632425 = weight(_text_:a in 2139) [ClassicSimilarity], result of:
          0.009632425 = score(doc=2139,freq=4.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.18016359 = fieldWeight in 2139, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.078125 = fieldNorm(doc=2139)
        0.007893822 = product of:
          0.015787644 = sum of:
            0.015787644 = weight(_text_:information in 2139) [ClassicSimilarity], result of:
              0.015787644 = score(doc=2139,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.19395474 = fieldWeight in 2139, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2139)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    For scientists, finding the information they want on the WWW is a hit-and-miss affair. But, as Declan Butler reports, more sophisticated and specialized search technlogies are promising to change all that
    Type
    a
  5. Garnsey, M.R.: What distance learners should know about information retrieval on the World Wide Web (2002) 0.01
    0.006548052 = product of:
      0.01637013 = sum of:
        0.005779455 = weight(_text_:a in 1626) [ClassicSimilarity], result of:
          0.005779455 = score(doc=1626,freq=4.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.10809815 = fieldWeight in 1626, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=1626)
        0.010590675 = product of:
          0.02118135 = sum of:
            0.02118135 = weight(_text_:information in 1626) [ClassicSimilarity], result of:
              0.02118135 = score(doc=1626,freq=10.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.2602176 = fieldWeight in 1626, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1626)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    The Internet can be a valuable tool allowing distance learners to access information not available locally. Search engines are the most common means of locating relevant information an the Internet, but to use them efficiently students should be taught the basics of searching and how to evaluate the results. This article briefly reviews how Search engines work, studies comparing Search engines, and criteria useful in evaluating the quality of returned Web pages. Research indicates there are statistical differences in the precision of Search engines, with AltaVista ranking high in several studies. When evaluating the quality of Web pages, standard criteria used in evaluating print resources is appropriate, as well as additional criteria which relate to the Web site itself. Giving distance learners training in how to use Search engines and how to evaluate the results will allow them to access relevant information efficiently while ensuring that it is of adequate quality.
    Footnote
    Part of an issue devoted to "Distance learning: information access and services for virtual users", publ. by Haworth Press
    Type
    a
  6. Sieverts, E.: Liever browsen dan zoeken (1998) 0.01
    0.0063011474 = product of:
      0.015752869 = sum of:
        0.009437811 = weight(_text_:a in 4722) [ClassicSimilarity], result of:
          0.009437811 = score(doc=4722,freq=6.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.17652355 = fieldWeight in 4722, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=4722)
        0.006315058 = product of:
          0.012630116 = sum of:
            0.012630116 = weight(_text_:information in 4722) [ClassicSimilarity], result of:
              0.012630116 = score(doc=4722,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.1551638 = fieldWeight in 4722, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4722)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Despite development of the WWW searchers still experience difficulties following links between sites and cannot be sure that a site contains the required information. 3 software programs developed to guide users through the maze of hyperlinks are: Dynamic diagrams, the Hyperbolic tree, and the Brain. in contrast to the other programs which operate on webservers and display hyperlinks in diagrammatic form the Brain is installed on individual PCs and can be customised to meet users' requirements
    Footnote
    Übers. d. Titels: A preference for browsing rather than searching
    Type
    a
  7. Granum, G.; Barker, P.: ¬An EASIER way to search online engineering resource (2000) 0.01
    0.0063011474 = product of:
      0.015752869 = sum of:
        0.009437811 = weight(_text_:a in 4876) [ClassicSimilarity], result of:
          0.009437811 = score(doc=4876,freq=6.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.17652355 = fieldWeight in 4876, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=4876)
        0.006315058 = product of:
          0.012630116 = sum of:
            0.012630116 = weight(_text_:information in 4876) [ClassicSimilarity], result of:
              0.012630116 = score(doc=4876,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.1551638 = fieldWeight in 4876, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4876)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    EEVL consists of several distinct resources, which exist as separate databases. This article describes the approach taken to tackle a particular problem that was identified through evaluation studies, namely, that searches of the EEVL catalogue too frequently matched nor records. The solution described in this paper is a cross-search facility for 3 of the EEVL databases
    Source
    Online information review. 24(2000) no.1, S.78-82
    Type
    a
  8. Nait-Baha, L.; Jackiewicz, A.; Djioua, B.; Laublet, P.: Query reformulation for information retrieval on the Web using the point of view methodology : preliminary results (2001) 0.01
    0.005948606 = product of:
      0.014871514 = sum of:
        0.008173384 = weight(_text_:a in 249) [ClassicSimilarity], result of:
          0.008173384 = score(doc=249,freq=8.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.15287387 = fieldWeight in 249, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=249)
        0.0066981306 = product of:
          0.013396261 = sum of:
            0.013396261 = weight(_text_:information in 249) [ClassicSimilarity], result of:
              0.013396261 = score(doc=249,freq=4.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.16457605 = fieldWeight in 249, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=249)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    The work we are presenting is devoted to the information collected on the WWW. By the term collected we mean the whole process of retrieving, extracting and presenting results to the user. This research is part of the RAP (Research, Analyze, Propose) project in which we propose to combine two methods: (i) query reformulation using linguistic markers according to a given point of view; and (ii) text semantic analysis by means of contextual exploration results (Descles, 1991). The general project architecture describing the interactions between the users, the RAP system and the WWW search engines is presented in Nait-Baha et al. (1998). We will focus this paper on showing how we use linguistic markers to reformulate the queries according to a given point of view
    Type
    a
  9. Gorbunov, A.L.: Relevance of Web documents : ghosts consensus method (2002) 0.01
    0.005948606 = product of:
      0.014871514 = sum of:
        0.008173384 = weight(_text_:a in 1005) [ClassicSimilarity], result of:
          0.008173384 = score(doc=1005,freq=8.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.15287387 = fieldWeight in 1005, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=1005)
        0.0066981306 = product of:
          0.013396261 = sum of:
            0.013396261 = weight(_text_:information in 1005) [ClassicSimilarity], result of:
              0.013396261 = score(doc=1005,freq=4.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.16457605 = fieldWeight in 1005, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1005)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    The dominant method currently used to improve the quality of Internet search systems is often called "digital democracy." Such an approach implies the utilization of the majority opinion of Internet users to determine the most relevant documents: for example, citation index usage for sorting of search results (google.com) or an enrichment of a query with terms that are asked frequently in relation with the query's theme. "Digital democracy" is an effective instrument in many cases, but it has an unavoidable shortcoming, which is a matter of principle: the average intellectual and cultural level of Internet users is very low- everyone knows what kind of information is dominant in Internet query statistics. Therefore, when one searches the Internet by means of "digital democracy" systems, one gets answers that reflect an underlying assumption that the user's mind potential is very low, and that his cultural interests are not demanding. Thus, it is more correct to use the term "digital ochlocracy" to refer to Internet search systems with "digital democracy." Based an the well-known mathematical mechanism of linear programming, we propose a method to solve the indicated problem.
    Source
    Journal of the American Society for Information Science and technology. 53(2002) no.10, S.783-788
    Type
    a
  10. Northern Light demonstrates Industry Search (1998) 0.01
    0.00588199 = product of:
      0.014704974 = sum of:
        0.0068111527 = weight(_text_:a in 2608) [ClassicSimilarity], result of:
          0.0068111527 = score(doc=2608,freq=2.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.12739488 = fieldWeight in 2608, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.078125 = fieldNorm(doc=2608)
        0.007893822 = product of:
          0.015787644 = sum of:
            0.015787644 = weight(_text_:information in 2608) [ClassicSimilarity], result of:
              0.015787644 = score(doc=2608,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.19395474 = fieldWeight in 2608, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2608)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Northern Light Technology have demonstrated Industry Search, an enhancement to its Northern Light service providing full text coverage of company and industry news, management practices and market research information via the WWW
    Type
    a
  11. Koenemann, J.; Lindner, H.-G.; Thomas, C.: Unternehmensportale : Von Suchmaschinen zum Wissensmanagement (2000) 0.01
    0.005735424 = product of:
      0.0143385595 = sum of:
        0.004767807 = weight(_text_:a in 5233) [ClassicSimilarity], result of:
          0.004767807 = score(doc=5233,freq=2.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.089176424 = fieldWeight in 5233, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5233)
        0.009570752 = product of:
          0.019141505 = sum of:
            0.019141505 = weight(_text_:information in 5233) [ClassicSimilarity], result of:
              0.019141505 = score(doc=5233,freq=6.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.23515764 = fieldWeight in 5233, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5233)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Aufgabe des Wissensmanagements ist es, den Mitarbeitern im Unternehmen entscheidungs- und handlungsrelevante Informationen bereitzustellen und die Mitarbeiter bei der intelligenten Verarbeitung dieser Informationen zu unterstützen. Ein hierzu genutztes Werkzeug von wachsender Bedeutung sind Unternehmensportale. Wir beschreiben kurz die Entwicklung von Portalen im World Wide Web (WWW), um dann Web-Portale von verschiedenen Arten von Unternehmensportalen abzugrenzen. Wir zeigen erwartete Funktionalitäten auf und stellen ein 5-Schichten Modell einer Gesamtarchitektur für Portale dar, welche die wesentlichen Komponenten umfasst. Im Anschluss werden die Besonderheiten der organisatorischen Realisierung und im Ausblick der Übergang von Portalen zum ,ubiquitous personalized information supply", der überall verfügbaren und individuellen Informationsversorgung behandelt
    Source
    nfd Information - Wissenschaft und Praxis. 51(2000) H.6, S.325-334
    Theme
    Information Resources Management
    Type
    a
  12. McMurdo, G.: How the Internet was indexed (1995) 0.01
    0.005593183 = product of:
      0.013982957 = sum of:
        0.005779455 = weight(_text_:a in 2411) [ClassicSimilarity], result of:
          0.005779455 = score(doc=2411,freq=4.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.10809815 = fieldWeight in 2411, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=2411)
        0.008203502 = product of:
          0.016407004 = sum of:
            0.016407004 = weight(_text_:information in 2411) [ClassicSimilarity], result of:
              0.016407004 = score(doc=2411,freq=6.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.20156369 = fieldWeight in 2411, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2411)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    The scope and characteristics of what may be considered the first three generations of automated Internet indexing systems are identified and described as to their methods of compiling their datasets, their search interfaces and the associated etymological metaphors and mythologies. These three are suggested to be: firstly, the Archie system for single keyword and regular expression searches of the file lists of anonymous ftp sites: secondly, the Veronica system for Boolean keyword in title searches of the world's gopher servers; thirdly, a range of software techniques jnown as robots and search engines, which compile searchable databases of information accessible via the WWW, such as the currently popular Lycos project at Carnegie Mellon University. The present dominance of WWW client software as the preferred interface to Internet information has led to provision of methods of also using the first two systems by this single interface, and these are also noted
    Source
    Journal of information science. 21(1995) no.6, S.479-489
    Type
    a
  13. Mansourian, I.: Web search efficacy : definition and implementation (2008) 0.01
    0.005431735 = product of:
      0.013579337 = sum of:
        0.009632425 = weight(_text_:a in 2565) [ClassicSimilarity], result of:
          0.009632425 = score(doc=2565,freq=16.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.18016359 = fieldWeight in 2565, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2565)
        0.003946911 = product of:
          0.007893822 = sum of:
            0.007893822 = weight(_text_:information in 2565) [ClassicSimilarity], result of:
              0.007893822 = score(doc=2565,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.09697737 = fieldWeight in 2565, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2565)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Purpose - This paper aims to report a number of factors that are perceived by web users as influential elements in their search procedure. The paper introduces a new conceptual measure called "web search efficacy" (hereafter WSE) to evaluate the performance of searches mainly based on users' perceptions. Design/methodology/approach - A rich dataset of a wider study was inductively re-explored to identify different categories that are perceived influential by web users on the final outcome of their searches. A selective review of the literature was carried out to discover to what extent previous research supports the findings of the current study. Findings - The analysis of the dataset led to the identification of five categories of influential factors. Within each group different factors have been recognized. Accordingly, the concept of WSE has been introduced. The five "Ss" which determine WSE are searcher's performance, search tool's performance, search strategy, search topic, and search situation. Research limitations/implications - The research body is scattered in different areas and it is difficult to carry out a comprehensive review. The WSE table, which is derived from the empirical data and was supported by previous research, can be employed for further research in various groups of web users. Originality/value - The paper contributes to the area of information seeking on the web by providing researchers with a new conceptual framework to evaluate the efficiency of each search session and identify the underlying factors on the final outcome of web searching.
    Type
    a
  14. Fu, T.; Abbasi, A.; Chen, H.: ¬A focused crawler for Dark Web forums (2010) 0.01
    0.005182888 = product of:
      0.012957219 = sum of:
        0.009010308 = weight(_text_:a in 3471) [ClassicSimilarity], result of:
          0.009010308 = score(doc=3471,freq=14.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.1685276 = fieldWeight in 3471, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3471)
        0.003946911 = product of:
          0.007893822 = sum of:
            0.007893822 = weight(_text_:information in 3471) [ClassicSimilarity], result of:
              0.007893822 = score(doc=3471,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.09697737 = fieldWeight in 3471, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3471)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    The unprecedented growth of the Internet has given rise to the Dark Web, the problematic facet of the Web associated with cybercrime, hate, and extremism. Despite the need for tools to collect and analyze Dark Web forums, the covert nature of this part of the Internet makes traditional Web crawling techniques insufficient for capturing such content. In this study, we propose a novel crawling system designed to collect Dark Web forum content. The system uses a human-assisted accessibility approach to gain access to Dark Web forums. Several URL ordering features and techniques enable efficient extraction of forum postings. The system also includes an incremental crawler coupled with a recall-improvement mechanism intended to facilitate enhanced retrieval and updating of collected content. Experiments conducted to evaluate the effectiveness of the human-assisted accessibility approach and the recall-improvement-based, incremental-update procedure yielded favorable results. The human-assisted approach significantly improved access to Dark Web forums while the incremental crawler with recall improvement also outperformed standard periodic- and incremental-update approaches. Using the system, we were able to collect over 100 Dark Web forums from three regions. A case study encompassing link and content analysis of collected forums was used to illustrate the value and importance of gathering and analyzing content from such online communities.
    Source
    Journal of the American Society for Information Science and Technology. 61(2010) no.6, S.1213-1231
    Type
    a
  15. Zins, C.: Models for classifying Internet resources (2002) 0.01
    0.0051638708 = product of:
      0.012909677 = sum of:
        0.008173384 = weight(_text_:a in 1160) [ClassicSimilarity], result of:
          0.008173384 = score(doc=1160,freq=8.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.15287387 = fieldWeight in 1160, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=1160)
        0.0047362936 = product of:
          0.009472587 = sum of:
            0.009472587 = weight(_text_:information in 1160) [ClassicSimilarity], result of:
              0.009472587 = score(doc=1160,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.116372846 = fieldWeight in 1160, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1160)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Designing systematic access to Internet resources is a major item an the agenda of researchers and practitioners in the field of information science, and is the focus of this study. A critical analysis of classification schemes used in major portals and Web classified directories exposes inconsistencies in the way they classify Internet resources. The inconsistencies indicate that the developers fall to differentiate the various classificatory models, and are unaware of their different rationales. The study establishes eight classificatory models for resources available to Internet users. Internet resources can be classified by subjects, objects, applications, users, locations, reference sources, media, and languages. The first five models are contentrelated; namely they characterize the content of the resource. The other three models are formst-related; namely they characterize the format of the resource or its technological infrastructure. The study identifies and formulates the eight classificatory models, analyzes their rationales, and discusses alternative ways to combine them in a faceted integrated classification scheme.
    Type
    a
  16. Lawrence, S.; Giles, C.L.: Accessibility and distribution of information on the Web (1999) 0.00
    0.0049910345 = product of:
      0.012477586 = sum of:
        0.005779455 = weight(_text_:a in 4952) [ClassicSimilarity], result of:
          0.005779455 = score(doc=4952,freq=4.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.10809815 = fieldWeight in 4952, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=4952)
        0.0066981306 = product of:
          0.013396261 = sum of:
            0.013396261 = weight(_text_:information in 4952) [ClassicSimilarity], result of:
              0.013396261 = score(doc=4952,freq=4.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.16457605 = fieldWeight in 4952, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4952)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Search engine coverage relative to the estimated size of the publicly indexable web has decreased substantially since December 97, with no engine indexing more than about 16% of the estimated size of the publicly indexable web. (Note that many queries can be satisfied with a relatively small database). Search engines are typically more likely to index sites that have more links to them (more 'popular' sites). They are also typically more likely to index US sites than non-US sites (AltaVista is an exception), and more likely to index commercial sites than educational sites. Indexing of new or modified pages byjust one of the major search engines can take months. 83% of sites contain commercial content and 6% contain scientific or educational content. Only 1.5% of sites contain pornographic content. The publicly indexable web contains an estimated 800 million pages as of February 1999, encompassing about 15 terabytes of information or about 6 terabytes of text after removing HTML tags, comments, and extra whitespace. The simple HTML "keywords" and "description" metatags are only used on the homepages of 34% of sites. Only 0.3% of sites use the Dublin Core metadata standard.
    Type
    a
  17. Spink, A.; Wolfram, D.; Jansen, B.J.; Saracevic, T.: Searching the Web : the public and their queries (2001) 0.00
    0.0049628555 = product of:
      0.012407139 = sum of:
        0.006092081 = weight(_text_:a in 6980) [ClassicSimilarity], result of:
          0.006092081 = score(doc=6980,freq=10.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.11394546 = fieldWeight in 6980, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.03125 = fieldNorm(doc=6980)
        0.006315058 = product of:
          0.012630116 = sum of:
            0.012630116 = weight(_text_:information in 6980) [ClassicSimilarity], result of:
              0.012630116 = score(doc=6980,freq=8.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.1551638 = fieldWeight in 6980, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.03125 = fieldNorm(doc=6980)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    In previous articles, we reported the state of Web searching in 1997 (Jansen, Spink, & Saracevic, 2000) and in 1999 (Spink, Wolfram, Jansen, & Saracevic, 2001). Such snapshot studies and statistics on Web use appear regularly (OCLC, 1999), but provide little information about Web searching trends. In this article, we compare and contrast results from our two previous studies of Excite queries' data sets, each containing over 1 million queries submitted by over 200,000 Excite users collected on 16 September 1997 and 20 December 1999. We examine how public Web searching changing during that 2-year time period. As Table 1 shows, the overall structure of Web queries in some areas did not change, while in others we see change from 1997 to 1999. Our comparison shows how Web searching changed incrementally and also dramatically. We see some moves toward greater simplicity, including shorter queries (i.e., fewer terms) and shorter sessions (i.e., fewer queries per user), with little modification (addition or deletion) of terms in subsequent queries. The trend toward shorter queries suggests that Web information content should target specific terms in order to reach Web users. Another trend was to view fewer pages of results per query. Most Excite users examined only one page of results per query, since an Excite results page contains ten ranked Web sites. Were users satisfied with the results and did not need to view more pages? It appears that the public continues to have a low tolerance of wading through retrieved sites. This decline in interactivity levels is a disturbing finding for the future of Web searching. Queries that included Boolean operators were in the minority, but the percentage increased between the two time periods. Most Boolean use involved the AND operator with many mistakes. The use of relevance feedback almost doubled from 1997 to 1999, but overall use was still small. An unusually large number of terms were used with low frequency, such as personal names, spelling errors, non-English words, and Web-specific terms, such as URLs. Web query vocabulary contains more words than found in large English texts in general. The public language of Web queries has its own and unique characteristics. How did Web searching topics change from 1997 to 1999? We classified a random sample of 2,414 queries from 1997 and 2,539 queries from 1999 into 11 categories (Table 2). From 1997 to 1999, Web searching shifted from entertainment, recreation and sex, and pornography, preferences to e-commerce-related topics under commerce, travel, employment, and economy. This shift coincided with changes in information distribution on the publicly indexed Web.
    Source
    Journal of the American Society for Information Science and technology. 52(2001) no.3, S.226-234
    Type
    a
  18. Jepsen, E.T.; Seiden, P.; Ingwersen, P.; Björneborn, L.; Borlund, P.: Characteristics of scientific Web publications : preliminary data gathering and analysis (2004) 0.00
    0.0049571716 = product of:
      0.012392929 = sum of:
        0.0068111527 = weight(_text_:a in 3091) [ClassicSimilarity], result of:
          0.0068111527 = score(doc=3091,freq=8.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.12739488 = fieldWeight in 3091, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3091)
        0.0055817757 = product of:
          0.011163551 = sum of:
            0.011163551 = weight(_text_:information in 3091) [ClassicSimilarity], result of:
              0.011163551 = score(doc=3091,freq=4.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.13714671 = fieldWeight in 3091, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3091)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Because of the increasing presence of scientific publications an the Web, combined with the existing difficulties in easily verifying and retrieving these publications, research an techniques and methods for retrieval of scientific Web publications is called for. In this article, we report an the initial steps taken toward the construction of a test collection of scientific Web publications within the subject domain of plant biology. The steps reported are those of data gathering and data analysis aiming at identifying characteristics of scientific Web publications. The data used in this article were generated based an specifically selected domain topics that are searched for in three publicly accessible search engines (Google, AlITheWeb, and AItaVista). A sample of the retrieved hits was analyzed with regard to how various publication attributes correlated with the scientific quality of the content and whether this information could be employed to harvest, filter, and rank Web publications. The attributes analyzed were inlinks, outlinks, bibliographic references, file format, language, search engine overlap, structural position (according to site structure), and the occurrence of various types of metadata. As could be expected, the ranked output differs between the three search engines. Apparently, this is caused by differences in ranking algorithms rather than the databases themselves. In fact, because scientific Web content in this subject domain receives few inlinks, both AItaVista and AlITheWeb retrieved a higher degree of accessible scientific content than Google. Because of the search engine cutoffs of accessible URLs, the feasibility of using search engine output for Web content analysis is also discussed.
    Source
    Journal of the American Society for Information Science and Technology. 55(2004) no.14, S.1239-1249
    Type
    a
  19. Ross, N.C.M.; Wolfram, D.: End user searching on the Internet : an analysis of term pair topics submitted to the Excite search engine (2000) 0.00
    0.004725861 = product of:
      0.011814652 = sum of:
        0.007078358 = weight(_text_:a in 4998) [ClassicSimilarity], result of:
          0.007078358 = score(doc=4998,freq=6.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.13239266 = fieldWeight in 4998, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=4998)
        0.0047362936 = product of:
          0.009472587 = sum of:
            0.009472587 = weight(_text_:information in 4998) [ClassicSimilarity], result of:
              0.009472587 = score(doc=4998,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.116372846 = fieldWeight in 4998, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4998)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Abstract
    Queries submitted to the Excite search engine were analyzed for subject content based on the cooccurrence of terms within multiterm queries. More than 1000 of the most frequently cooccurring term pairs were categorized into one or more of 30 developed subject areas. Subject area frequencies and their cooccurrences with one another were tallied and analyzed using hierarchical cluster analysis and multidimensional scaling. The cluster analyses revealed several anticipated and a few unanticipated groupings of subjects, resulting in several well-defined high-level clusters of broad subject areas. Multidimensional scaling of subject cooccurrences revealed similar relationships among the different subject categories. Applications that arise from a better understanding of the topics users search and their relationships are discussed
    Source
    Journal of the American Society for Information Science. 51(2000) no.10, S.949-958
    Type
    a
  20. Kennedy, S.D.: How to find subjects and subject experts (1996) 0.00
    0.0047055925 = product of:
      0.011763981 = sum of:
        0.005448922 = weight(_text_:a in 4531) [ClassicSimilarity], result of:
          0.005448922 = score(doc=4531,freq=2.0), product of:
            0.053464882 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046368346 = queryNorm
            0.10191591 = fieldWeight in 4531, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=4531)
        0.006315058 = product of:
          0.012630116 = sum of:
            0.012630116 = weight(_text_:information in 4531) [ClassicSimilarity], result of:
              0.012630116 = score(doc=4531,freq=2.0), product of:
                0.08139861 = queryWeight, product of:
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.046368346 = queryNorm
                0.1551638 = fieldWeight in 4531, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  1.7554779 = idf(docFreq=20772, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4531)
          0.5 = coord(1/2)
      0.4 = coord(2/5)
    
    Source
    Information today. 13(1996) no.2, S.46-47
    Type
    a

Languages

  • e 41
  • d 27
  • nl 1
  • More… Less…