Search (310 results, page 1 of 16)

  • × theme_ss:"Suchmaschinen"
  1. Li, L.; Shang, Y.; Zhang, W.: Improvement of HITS-based algorithms on Web documents 0.05
    0.04980251 = product of:
      0.24901254 = sum of:
        0.24901254 = weight(_text_:3a in 2514) [ClassicSimilarity], result of:
          0.24901254 = score(doc=2514,freq=2.0), product of:
            0.4430686 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.052260913 = queryNorm
            0.56201804 = fieldWeight in 2514, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=2514)
      0.2 = coord(1/5)
    
    Content
    Vgl.: http%3A%2F%2Fdelab.csd.auth.gr%2F~dimitris%2Fcourses%2Fir_spring06%2Fpage_rank_computing%2Fp527-li.pdf. Vgl. auch: http://www2002.org/CDROM/refereed/643/.
  2. Marchiori, M.: ¬The quest for correct information on the Web : hyper search engines (1997) 0.04
    0.043250576 = product of:
      0.10812644 = sum of:
        0.05856201 = weight(_text_:it in 7453) [ClassicSimilarity], result of:
          0.05856201 = score(doc=7453,freq=6.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.38743722 = fieldWeight in 7453, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.0546875 = fieldNorm(doc=7453)
        0.04956443 = weight(_text_:22 in 7453) [ClassicSimilarity], result of:
          0.04956443 = score(doc=7453,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.2708308 = fieldWeight in 7453, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0546875 = fieldNorm(doc=7453)
      0.4 = coord(2/5)
    
    Abstract
    Presents a novel method to extract from a web object its hyper informative content, in contrast with current search engines, which only deal with the textual information content. This method is not only valuable per se, but it is shown to be able to considerably increase the precision of current search engines. It integrates with existing search engine technology since it can be implemented on top of every search engine, acting as a post-processor, thus automatically transforming a search engine into its corresponding hyper version. Shows how the hyper information can be usefully employed to face the search engines persuasion problem
    Date
    1. 8.1996 22:08:06
  3. Bensman, S.J.: Eugene Garfield, Francis Narin, and PageRank : the theoretical bases of the Google search engine (2013) 0.04
    0.038114388 = product of:
      0.09528597 = sum of:
        0.038640905 = weight(_text_:it in 1149) [ClassicSimilarity], result of:
          0.038640905 = score(doc=1149,freq=2.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.25564227 = fieldWeight in 1149, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.0625 = fieldNorm(doc=1149)
        0.05664506 = weight(_text_:22 in 1149) [ClassicSimilarity], result of:
          0.05664506 = score(doc=1149,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.30952093 = fieldWeight in 1149, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0625 = fieldNorm(doc=1149)
      0.4 = coord(2/5)
    
    Abstract
    This paper presents a test of the validity of using Google Scholar to evaluate the publications of researchers by comparing the premises on which its search engine, PageRank, is based, to those of Garfield's theory of citation indexing. It finds that the premises are identical and that PageRank and Garfield's theory of citation indexing validate each other.
    Date
    17.12.2013 11:02:22
  4. Chen, L.-C.: Next generation search engine for the result clustering technology (2012) 0.04
    0.03707192 = product of:
      0.0926798 = sum of:
        0.050196007 = weight(_text_:it in 105) [ClassicSimilarity], result of:
          0.050196007 = score(doc=105,freq=6.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.33208904 = fieldWeight in 105, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.046875 = fieldNorm(doc=105)
        0.042483795 = weight(_text_:22 in 105) [ClassicSimilarity], result of:
          0.042483795 = score(doc=105,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.23214069 = fieldWeight in 105, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.046875 = fieldNorm(doc=105)
      0.4 = coord(2/5)
    
    Abstract
    Result clustering has recently attracted a lot of attention to provide the users with a succinct overview of relevant search results than traditional search engines. This chapter proposes a mixed clustering method to organize all returned search results into a hierarchical tree structure. The clustering method accomplishes two main tasks, one is label construction and the other is tree building. This chapter uses precision to measure the quality of clustering results. According to the results of experiments, the author preliminarily concluded that the performance of the system is better than many other well-known commercial and academic systems. This chapter makes several contributions. First, it presents a high performance system based on the clustering method. Second, it develops a divisive hierarchical clustering algorithm to organize all returned snippets into hierarchical tree structure. Third, it performs a wide range of experimental analyses to show that almost all commercial systems are significantly better than most current academic systems.
    Date
    17. 4.2012 15:22:11
  5. Schultheiß, G.F.: Google, Goggle, Google, ... : Whose Mind is it Anywhere? Identifying and Meeting Divers User Needs in the Ongoing Sattle for Mindshare - NFAIS 47th Annual Conference, Philadelphia, USA vom 27. Februar bis 1. März 2005 (2005) 0.03
    0.033387464 = product of:
      0.08346866 = sum of:
        0.04098487 = weight(_text_:it in 3421) [ClassicSimilarity], result of:
          0.04098487 = score(doc=3421,freq=4.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.27114958 = fieldWeight in 3421, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.046875 = fieldNorm(doc=3421)
        0.042483795 = weight(_text_:22 in 3421) [ClassicSimilarity], result of:
          0.042483795 = score(doc=3421,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.23214069 = fieldWeight in 3421, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.046875 = fieldNorm(doc=3421)
      0.4 = coord(2/5)
    
    Abstract
    Das Herausfinden und Lösen verschiedener Nutzerbedürfnisse als Hauptthema der Konferenz warf gleich mehrere Fragen auf: Erstens, wer sind die Nutzer? Zweitens, was wollen sie? Und wer beurteilt das und aus welcher Sicht? Welche Auswirkungen haben die realisierten Maßnahmen? Die etwa 200 Teilnehmer aus neun Nationen, davon 21 aus europäischen Ländern, davon drei aus Deutschland, wurden in sechs Vortragsblöcken mit einer Vielzahl von Aspekten konfrontiert, die wie so oft in unserer rasanten Entwicklung im IT-Bereich nur Lösungsansätze aufzeigten oder die Schwierigkeiten mit den Angeboten. Herausragendes Merkmal war der mehrfache Bezug auf die Marke "Google"!
    Date
    22. 5.2005 13:38:26
  6. Fluhr, C.: Crosslingual access to photo databases (2012) 0.03
    0.033387464 = product of:
      0.08346866 = sum of:
        0.04098487 = weight(_text_:it in 93) [ClassicSimilarity], result of:
          0.04098487 = score(doc=93,freq=4.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.27114958 = fieldWeight in 93, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.046875 = fieldNorm(doc=93)
        0.042483795 = weight(_text_:22 in 93) [ClassicSimilarity], result of:
          0.042483795 = score(doc=93,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.23214069 = fieldWeight in 93, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.046875 = fieldNorm(doc=93)
      0.4 = coord(2/5)
    
    Abstract
    This paper is about search of photos in photo databases of agencies which sell photos over the Internet. The problem is far from the behavior of photo databases managed by librarians and also far from the corpora generally used for research purposes. The descriptions use mainly single words and it is well known that it is not the best way to have a good search. This increases the problem of semantic ambiguity. This problem of semantic ambiguity is crucial for cross-language querying. On the other hand, users are not aware of documentation techniques and use generally very simple queries but want to get precise answers. This paper gives the experience gained in a 3 year use (2006-2008) of a cross-language access to several of the main international commercial photo databases. The languages used were French, English, and German.
    Date
    17. 4.2012 14:25:22
  7. Lawrence, S.; Giles, C.L.: Inquirus, the NECI meta search engine (1998) 0.03
    0.033350088 = product of:
      0.083375216 = sum of:
        0.03381079 = weight(_text_:it in 3604) [ClassicSimilarity], result of:
          0.03381079 = score(doc=3604,freq=2.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.22368698 = fieldWeight in 3604, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3604)
        0.04956443 = weight(_text_:22 in 3604) [ClassicSimilarity], result of:
          0.04956443 = score(doc=3604,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.2708308 = fieldWeight in 3604, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3604)
      0.4 = coord(2/5)
    
    Abstract
    Presents Inquirus, a WWW meta search engine which works by downloading and analysing the individual documents. It makes improvements over existing search engines in a number of areas: more useful document summaries incorporating query term context, identification of both pages which no longer exist and pages which no longer contain the query terms, advanced detection of duplicate pages, improved document ranking using proximity information, dramatically improved precision for certain queries by using specific expressive forms, and quick jump links and highlighting when viewing the full document
    Date
    1. 8.1996 22:08:06
  8. Jenkins, C.: Automatic classification of Web resources using Java and Dewey Decimal Classification (1998) 0.03
    0.033350088 = product of:
      0.083375216 = sum of:
        0.03381079 = weight(_text_:it in 1673) [ClassicSimilarity], result of:
          0.03381079 = score(doc=1673,freq=2.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.22368698 = fieldWeight in 1673, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1673)
        0.04956443 = weight(_text_:22 in 1673) [ClassicSimilarity], result of:
          0.04956443 = score(doc=1673,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.2708308 = fieldWeight in 1673, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1673)
      0.4 = coord(2/5)
    
    Abstract
    The Wolverhampton Web Library (WWLib) is a WWW search engine that provides access to UK based information. The experimental version developed in 1995, was a success but highlighted the need for a much higher degree of automation. An interesting feature of the experimental WWLib was that it organised information according to DDC. Discusses the advantages of classification and describes the automatic classifier that is being developed in Java as part of the new, fully automated WWLib
    Date
    1. 8.1996 22:08:06
  9. Aloteibi, S.; Sanderson, M.: Analyzing geographic query reformulation : an exploratory study (2014) 0.03
    0.030893266 = product of:
      0.077233166 = sum of:
        0.041830003 = weight(_text_:it in 1177) [ClassicSimilarity], result of:
          0.041830003 = score(doc=1177,freq=6.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.27674085 = fieldWeight in 1177, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1177)
        0.035403162 = weight(_text_:22 in 1177) [ClassicSimilarity], result of:
          0.035403162 = score(doc=1177,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.19345059 = fieldWeight in 1177, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1177)
      0.4 = coord(2/5)
    
    Abstract
    Search engine users typically engage in multiquery sessions in their quest to fulfill their information needs. Despite a plethora of research findings suggesting that a significant group of users look for information within a specific geographical scope, existing reformulation studies lack a focused analysis of how users reformulate geographic queries. This study comprehensively investigates the ways in which users reformulate such needs in an attempt to fill this gap in the literature. Reformulated sessions were sampled from a query log of a major search engine to extract 2,400 entries that were manually inspected to filter geo sessions. This filter identified 471 search sessions that included geographical intent, and these sessions were analyzed quantitatively and qualitatively. The results revealed that one in five of the users who reformulated their queries were looking for geographically related information. They reformulated their queries by changing the content of the query rather than the structure. Users were not following a unified sequence of modifications and instead performed a single reformulation action. However, in some cases it was possible to anticipate their next move. A number of tasks in geo modifications were identified, including standard, multi-needs, multi-places, and hybrid approaches. The research concludes that it is important to specialize query reformulation studies to focus on particular query types rather than generically analyzing them, as it is apparent that geographic queries have their special reformulation characteristics.
    Date
    26. 1.2014 18:48:22
  10. Duval, B.K.; Main, L.: Searching the Internet : part 2 trail-blazers (1997) 0.03
    0.02858579 = product of:
      0.07146447 = sum of:
        0.028980678 = weight(_text_:it in 858) [ClassicSimilarity], result of:
          0.028980678 = score(doc=858,freq=2.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.19173169 = fieldWeight in 858, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.046875 = fieldNorm(doc=858)
        0.042483795 = weight(_text_:22 in 858) [ClassicSimilarity], result of:
          0.042483795 = score(doc=858,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.23214069 = fieldWeight in 858, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.046875 = fieldNorm(doc=858)
      0.4 = coord(2/5)
    
    Abstract
    Presents a guide to searching for information on the Internet covering Research-It; familiar quotations: a collection of passages, phrases and proverbs traced to their sources in ancient and modern literature by John Bartlett; the Internet Public Library Reference Center; SearchERIC Database; Britannica Online; Britannica's Lives; The complete works of William Shakespeare; Flicks/Movie Schedules and Reviews; the Electronic Newsstand; CNN Interactive; Time Warner's Pathfinder; Electronic Newspapers from all 50 States; Yahoo, News; Newspapers; Techweb; ZDNet; the On-line Books Page; Columbia University Bartleby Library; the Children's Literature Web Guide; National Institutes of Health; US Census Bureau; Earthquake Info; US Postal Service Zip+4 Lookup; the Federal Web Locator; World Wide Web Virtual Library; US Government Information Sources; Index of the Constitution of the US; US States Code; Find California Code; Dearch for Bills; California Tenant's Rights; The Online Career Center; QuickAID Home Page; City.Net; Netscape's Destinations Button; International Telephone Directory; World Alumni Net; Archives of Adoptees and Birth Parents; and World Wide Registry Matching Adoptees with Birth Parents
    Date
    6. 3.1997 16:22:15
  11. Carrière, S.J.; Kazman, R.: Webquery : searching and visualising the Web through connectivity (1997) 0.03
    0.02858579 = product of:
      0.07146447 = sum of:
        0.028980678 = weight(_text_:it in 2674) [ClassicSimilarity], result of:
          0.028980678 = score(doc=2674,freq=2.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.19173169 = fieldWeight in 2674, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.046875 = fieldNorm(doc=2674)
        0.042483795 = weight(_text_:22 in 2674) [ClassicSimilarity], result of:
          0.042483795 = score(doc=2674,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.23214069 = fieldWeight in 2674, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.046875 = fieldNorm(doc=2674)
      0.4 = coord(2/5)
    
    Abstract
    The WebQuery system offers a powerful new method for searching the Web based on connectivity and content. Examines links among the nodes returned in a keyword-based query. Rankes the nodes, giving the highest rank to the most highly connected nodes. By doing so, finds hot spots on the Web that contain information germane to a user's query. WebQuery not only ranks and filters the results of a Web query; it also extends the result set beyond what the search engine retrieves, by finding interesting sites that are highly connected to those sites returned by the original query. Even with WebQuery filering and ranking query results, the result set can be enormous. Explores techniques for visualizing the returned information and discusses the criteria for using each of the technique
    Date
    1. 8.1996 22:08:06
  12. Nicholson, S.; Sierra, T.; Eseryel, U.Y.; Park, J.-H.; Barkow, P.; Pozo, E.J.; Ward, J.: How much of it is real? : analysis of paid placement in Web search engine results (2006) 0.03
    0.02858579 = product of:
      0.07146447 = sum of:
        0.028980678 = weight(_text_:it in 5278) [ClassicSimilarity], result of:
          0.028980678 = score(doc=5278,freq=2.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.19173169 = fieldWeight in 5278, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.046875 = fieldNorm(doc=5278)
        0.042483795 = weight(_text_:22 in 5278) [ClassicSimilarity], result of:
          0.042483795 = score(doc=5278,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.23214069 = fieldWeight in 5278, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.046875 = fieldNorm(doc=5278)
      0.4 = coord(2/5)
    
    Date
    22. 7.2006 16:32:57
  13. Golderman, G.M.; Connolly, B.: Between the book covers : going beyond OPAC keyword searching with the deep linking capabilities of Google Scholar and Google Book Search (2004/05) 0.03
    0.027822888 = product of:
      0.06955722 = sum of:
        0.034154054 = weight(_text_:it in 731) [ClassicSimilarity], result of:
          0.034154054 = score(doc=731,freq=4.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.22595796 = fieldWeight in 731, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.0390625 = fieldNorm(doc=731)
        0.035403162 = weight(_text_:22 in 731) [ClassicSimilarity], result of:
          0.035403162 = score(doc=731,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.19345059 = fieldWeight in 731, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0390625 = fieldNorm(doc=731)
      0.4 = coord(2/5)
    
    Abstract
    One finding of the 2006 OCLC study of College Students' Perceptions of Libraries and Information Resources was that students expressed equal levels of trust in libraries and search engines when it came to meeting their information needs in a way that they felt was authoritative. Seeking to incorporate this insight into our own instructional methodology, Schaffer Library at Union College has attempted to engineer a shift from Google to Google Scholar among our student users by representing Scholar as a viable adjunct to the catalog and to snore traditional electronic resources. By attempting to engage student researchers on their own terms, we have discovered that most of them react enthusiastically to the revelation that the Google they think they know so well is, it turns out, a multifaceted resource that is capable of delivering the sort of scholarly information that will meet with their professors' approval. Specifically, this article focuses on the fact that many Google Scholar searches link hack to our own Web catalog where they identify useful book titles that direct OPAC keyword searches have missed.
    Date
    2.12.2007 19:39:22
  14. Alqaraleh, S.; Ramadan, O.; Salamah, M.: Efficient watcher based web crawler design (2015) 0.03
    0.027822888 = product of:
      0.06955722 = sum of:
        0.034154054 = weight(_text_:it in 1627) [ClassicSimilarity], result of:
          0.034154054 = score(doc=1627,freq=4.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.22595796 = fieldWeight in 1627, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1627)
        0.035403162 = weight(_text_:22 in 1627) [ClassicSimilarity], result of:
          0.035403162 = score(doc=1627,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.19345059 = fieldWeight in 1627, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1627)
      0.4 = coord(2/5)
    
    Abstract
    Purpose The purpose of this paper is to design a watcher-based crawler (WBC) that has the ability of crawling static and dynamic web sites, and can download only the updated and newly added web pages. Design/methodology/approach In the proposed WBC crawler, a watcher file, which can be uploaded to the web sites servers, prepares a report that contains the addresses of the updated and the newly added web pages. In addition, the WBC is split into five units, where each unit is responsible for performing a specific crawling process. Findings Several experiments have been conducted and it has been observed that the proposed WBC increases the number of uniquely visited static and dynamic web sites as compared with the existing crawling techniques. In addition, the proposed watcher file not only allows the crawlers to visit the updated and newly web pages, but also solves the crawlers overlapping and communication problems. Originality/value The proposed WBC performs all crawling processes in the sense that it detects all updated and newly added pages automatically without any human explicit intervention or downloading the entire web sites.
    Date
    20. 1.2015 18:30:22
  15. Drabenstott, K.M.: Web search strategies (2000) 0.02
    0.024714613 = product of:
      0.061786532 = sum of:
        0.033464003 = weight(_text_:it in 1188) [ClassicSimilarity], result of:
          0.033464003 = score(doc=1188,freq=6.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.22139269 = fieldWeight in 1188, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.03125 = fieldNorm(doc=1188)
        0.02832253 = weight(_text_:22 in 1188) [ClassicSimilarity], result of:
          0.02832253 = score(doc=1188,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.15476047 = fieldWeight in 1188, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.03125 = fieldNorm(doc=1188)
      0.4 = coord(2/5)
    
    Abstract
    Surfing the World Wide Web used to be cool, dude, real cool. But things have gotten hot - so hot that finding something useful an the Web is no longer cool. It is suffocating Web searchers in the smoke and debris of mountain-sized lists of hits, decisions about which search engines they should use, whether they will get lost in the dizzying maze of a subject directory, use the right syntax for the search engine at hand, enter keywords that are likely to retrieve hits an the topics they have in mind, or enlist a browser that has sufficient functionality to display the most promising hits. When it comes to Web searching, in a few short years we have gone from the cool image of surfing the Web into the frying pan of searching the Web. We can turn down the heat by rethinking what Web searchers are doing and introduce some order into the chaos. Web search strategies that are tool-based-oriented to specific Web searching tools such as search en gines, subject directories, and meta search engines-have been widely promoted, and these strategies are just not working. It is time to dissect what Web searching tools expect from searchers and adjust our search strategies to these new tools. This discussion offers Web searchers help in the form of search strategies that are based an strategies that librarians have been using for a long time to search commercial information retrieval systems like Dialog, NEXIS, Wilsonline, FirstSearch, and Data-Star.
    Date
    22. 9.1997 19:16:05
  16. Großjohann, K.: Gathering-, Harvesting-, Suchmaschinen (1996) 0.02
    0.024032464 = product of:
      0.12016232 = sum of:
        0.12016232 = weight(_text_:22 in 3227) [ClassicSimilarity], result of:
          0.12016232 = score(doc=3227,freq=4.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.6565931 = fieldWeight in 3227, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.09375 = fieldNorm(doc=3227)
      0.2 = coord(1/5)
    
    Date
    7. 2.1996 22:38:41
    Pages
    22 S
  17. Höfer, W.: Detektive im Web (1999) 0.02
    0.024032464 = product of:
      0.12016232 = sum of:
        0.12016232 = weight(_text_:22 in 4007) [ClassicSimilarity], result of:
          0.12016232 = score(doc=4007,freq=4.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.6565931 = fieldWeight in 4007, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.09375 = fieldNorm(doc=4007)
      0.2 = coord(1/5)
    
    Date
    22. 8.1999 20:22:06
  18. Rensman, J.: Blick ins Getriebe (1999) 0.02
    0.024032464 = product of:
      0.12016232 = sum of:
        0.12016232 = weight(_text_:22 in 4009) [ClassicSimilarity], result of:
          0.12016232 = score(doc=4009,freq=4.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.6565931 = fieldWeight in 4009, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.09375 = fieldNorm(doc=4009)
      0.2 = coord(1/5)
    
    Date
    22. 8.1999 21:22:59
  19. Baeza-Yates, R.; Boldi, P.; Castillo, C.: Generalizing PageRank : damping functions for linkbased ranking algorithms (2006) 0.02
    0.023821492 = product of:
      0.059553728 = sum of:
        0.024150565 = weight(_text_:it in 2565) [ClassicSimilarity], result of:
          0.024150565 = score(doc=2565,freq=2.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.15977642 = fieldWeight in 2565, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2565)
        0.035403162 = weight(_text_:22 in 2565) [ClassicSimilarity], result of:
          0.035403162 = score(doc=2565,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.19345059 = fieldWeight in 2565, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2565)
      0.4 = coord(2/5)
    
    Abstract
    This paper introduces a family of link-based ranking algorithms that propagate page importance through links. In these algorithms there is a damping function that decreases with distance, so a direct link implies more endorsement than a link through a long path. PageRank is the most widely known ranking function of this family. The main objective of this paper is to determine whether this family of ranking techniques has some interest per se, and how different choices for the damping function impact on rank quality and on convergence speed. Even though our results suggest that PageRank can be approximated with other simpler forms of rankings that may be computed more efficiently, our focus is of more speculative nature, in that it aims at separating the kernel of PageRank, that is, link-based importance propagation, from the way propagation decays over paths. We focus on three damping functions, having linear, exponential, and hyperbolic decay on the lengths of the paths. The exponential decay corresponds to PageRank, and the other functions are new. Our presentation includes algorithms, analysis, comparisons and experiments that study their behavior under different parameters in real Web graph data. Among other results, we show how to calculate a linear approximation that induces a page ordering that is almost identical to PageRank's using a fixed small number of iterations; comparisons were performed using Kendall's tau on large domain datasets.
    Date
    16. 1.2016 10:22:28
  20. Lewandowski, D.; Sünkler, S.: What does Google recommend when you want to compare insurance offerings? (2019) 0.02
    0.023821492 = product of:
      0.059553728 = sum of:
        0.024150565 = weight(_text_:it in 5288) [ClassicSimilarity], result of:
          0.024150565 = score(doc=5288,freq=2.0), product of:
            0.15115225 = queryWeight, product of:
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.052260913 = queryNorm
            0.15977642 = fieldWeight in 5288, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.892262 = idf(docFreq=6664, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5288)
        0.035403162 = weight(_text_:22 in 5288) [ClassicSimilarity], result of:
          0.035403162 = score(doc=5288,freq=2.0), product of:
            0.18300882 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.052260913 = queryNorm
            0.19345059 = fieldWeight in 5288, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5288)
      0.4 = coord(2/5)
    
    Abstract
    Purpose The purpose of this paper is to describe a new method to improve the analysis of search engine results by considering the provider level as well as the domain level. This approach is tested by conducting a study using queries on the topic of insurance comparisons. Design/methodology/approach The authors conducted an empirical study that analyses the results of search queries aimed at comparing insurance companies. The authors used a self-developed software system that automatically queries commercial search engines and automatically extracts the content of the returned result pages for further data analysis. The data analysis was carried out using the KNIME Analytics Platform. Findings Google's top search results are served by only a few providers that frequently appear in these results. The authors show that some providers operate several domains on the same topic and that these domains appear for the same queries in the result lists. Research limitations/implications The authors demonstrate the feasibility of this approach and draw conclusions for further investigations from the empirical study. However, the study is a limited use case based on a limited number of search queries. Originality/value The proposed method allows large-scale analysis of the composition of the top results from commercial search engines. It allows using valid empirical data to determine what users actually see on the search engine result pages.
    Date
    20. 1.2015 18:30:22

Years

Languages

  • e 211
  • d 94
  • f 2
  • nl 2
  • More… Less…

Types

  • a 264
  • el 35
  • m 19
  • s 3
  • x 3
  • p 2
  • r 1
  • More… Less…

Subjects