Search (281 results, page 2 of 15)

  • × language_ss:"e"
  • × theme_ss:"Suchmaschinen"
  • × year_i:[2000 TO 2010}
  1. Fischer, T.; Neuroth, H.: SSG-FI - special subject gateways to high quality Internet resources for scientific users (2000) 0.02
    0.015996836 = product of:
      0.023995254 = sum of:
        0.0056313644 = weight(_text_:a in 4873) [ClassicSimilarity], result of:
          0.0056313644 = score(doc=4873,freq=4.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.10809815 = fieldWeight in 4873, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=4873)
        0.01836389 = product of:
          0.03672778 = sum of:
            0.03672778 = weight(_text_:22 in 4873) [ClassicSimilarity], result of:
              0.03672778 = score(doc=4873,freq=2.0), product of:
                0.15821345 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045180224 = queryNorm
                0.23214069 = fieldWeight in 4873, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4873)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Project SSG-FI at SUB Göttingen provides special subject gateways to international high quality Internet resources for scientific users. Internet sites are selected by subject specialists and described using an extension of qualified Dublin Core metadata. A basic evaluation is added. These descriptions are freely available and can be searched and browsed. These are now subject gateways for 3 subject ares: earth sciences (GeoGuide); mathematics (MathGuide); and Anglo-American culture (split into HistoryGuide and AnglistikGuide). Together they receive about 3.300 'hard' requests per day, thus reaching over 1 million requests per year. The project SSG-FI behind these guides is open to collaboration. Institutions and private persons wishing to contribute can notify the SSG-FI team or send full data sets. Regular contributors can request registration with the project to access the database via the Internet and create and edit records
    Date
    22. 6.2002 19:40:42
    Type
    a
  2. Large, A.; Beheshti, J.; Rahman, T.: Design criteria for children's Web portals : the users speak out (2002) 0.02
    0.015996836 = product of:
      0.023995254 = sum of:
        0.0056313644 = weight(_text_:a in 197) [ClassicSimilarity], result of:
          0.0056313644 = score(doc=197,freq=4.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.10809815 = fieldWeight in 197, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=197)
        0.01836389 = product of:
          0.03672778 = sum of:
            0.03672778 = weight(_text_:22 in 197) [ClassicSimilarity], result of:
              0.03672778 = score(doc=197,freq=2.0), product of:
                0.15821345 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045180224 = queryNorm
                0.23214069 = fieldWeight in 197, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=197)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Date
    2. 6.2005 10:34:22
    Type
    a
  3. Golderman, G.M.; Connolly, B.: Between the book covers : going beyond OPAC keyword searching with the deep linking capabilities of Google Scholar and Google Book Search (2004/05) 0.02
    0.015148811 = product of:
      0.022723217 = sum of:
        0.0074199745 = weight(_text_:a in 731) [ClassicSimilarity], result of:
          0.0074199745 = score(doc=731,freq=10.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.14243183 = fieldWeight in 731, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=731)
        0.015303242 = product of:
          0.030606484 = sum of:
            0.030606484 = weight(_text_:22 in 731) [ClassicSimilarity], result of:
              0.030606484 = score(doc=731,freq=2.0), product of:
                0.15821345 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045180224 = queryNorm
                0.19345059 = fieldWeight in 731, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=731)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    One finding of the 2006 OCLC study of College Students' Perceptions of Libraries and Information Resources was that students expressed equal levels of trust in libraries and search engines when it came to meeting their information needs in a way that they felt was authoritative. Seeking to incorporate this insight into our own instructional methodology, Schaffer Library at Union College has attempted to engineer a shift from Google to Google Scholar among our student users by representing Scholar as a viable adjunct to the catalog and to snore traditional electronic resources. By attempting to engage student researchers on their own terms, we have discovered that most of them react enthusiastically to the revelation that the Google they think they know so well is, it turns out, a multifaceted resource that is capable of delivering the sort of scholarly information that will meet with their professors' approval. Specifically, this article focuses on the fact that many Google Scholar searches link hack to our own Web catalog where they identify useful book titles that direct OPAC keyword searches have missed.
    Date
    2.12.2007 19:39:22
    Type
    a
  4. Loia, V.; Pedrycz, W.; Senatore, S.; Sessa, M.I.: Web navigation support by means of proximity-driven assistant agents (2006) 0.01
    0.014626579 = product of:
      0.021939868 = sum of:
        0.0066366266 = weight(_text_:a in 5283) [ClassicSimilarity], result of:
          0.0066366266 = score(doc=5283,freq=8.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.12739488 = fieldWeight in 5283, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5283)
        0.015303242 = product of:
          0.030606484 = sum of:
            0.030606484 = weight(_text_:22 in 5283) [ClassicSimilarity], result of:
              0.030606484 = score(doc=5283,freq=2.0), product of:
                0.15821345 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045180224 = queryNorm
                0.19345059 = fieldWeight in 5283, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5283)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The explosive growth of the Web and the consequent exigency of the Web personalization domain have gained a key position in the direction of customization of the Web information to the needs of specific users, taking advantage of the knowledge acquired from the analysis of the user's navigational behavior (usage data) in correlation with other information collected in the Web context, namely, structure, content, and user profile data. This work presents an agent-based framework designed to help a user in achieving personalized navigation, by recommending related documents according to the user's responses in similar-pages searching mode. Our agent-based approach is grounded in the integration of different techniques and methodologies into a unique platform featuring user profiling, fuzzy multisets, proximity-oriented fuzzy clustering, and knowledge-based discovery technologies. Each of these methodologies serves to solve one facet of the general problem (discovering documents relevant to the user by searching the Web) and is treated by specialized agents that ultimately achieve the final functionality through cooperation and task distribution.
    Date
    22. 7.2006 16:59:13
    Type
    a
  5. Su, L.T.: ¬A comprehensive and systematic model of user evaluation of Web search engines : Il. An evaluation by undergraduates (2003) 0.01
    0.01403382 = product of:
      0.021050729 = sum of:
        0.005747488 = weight(_text_:a in 2117) [ClassicSimilarity], result of:
          0.005747488 = score(doc=2117,freq=6.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.11032722 = fieldWeight in 2117, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2117)
        0.015303242 = product of:
          0.030606484 = sum of:
            0.030606484 = weight(_text_:22 in 2117) [ClassicSimilarity], result of:
              0.030606484 = score(doc=2117,freq=2.0), product of:
                0.15821345 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045180224 = queryNorm
                0.19345059 = fieldWeight in 2117, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2117)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This paper presents an application of the model described in Part I to the evaluation of Web search engines by undergraduates. The study observed how 36 undergraduate used four major search engines to find information for their own individual problems and how they evaluated these engines based an actual interaction with the search engines. User evaluation was based an 16 performance measures representing five evaluation criteria: relevance, efficiency, utility, user satisfaction, and connectivity. Non-performance (user-related) measures were also applied. Each participant searched his/ her own topic an all four engines and provided satisfaction ratings for system features and interaction and reasons for satisfaction. Each also made relevance judgements of retrieved items in relation to his/her own information need and participated in post-search Interviews to provide reactions to the search results and overall performance. The study found significant differences in precision PR1 relative recall, user satisfaction with output display, time saving, value of search results, and overall performance among the four engines and also significant engine by discipline interactions an all these measures. In addition, the study found significant differences in user satisfaction with response time among four engines, and significant engine by discipline interaction in user satisfaction with search interface. None of the four search engines dominated in every aspect of the multidimensional evaluation. Content analysis of verbal data identified a number of user criteria and users evaluative comments based an these criteria. Results from both quantitative analysis and content analysis provide insight for system design and development, and useful feedback an strengths and weaknesses of search engines for system improvement
    Date
    24. 1.2004 18:27:22
    Type
    a
  6. Drabenstott, K.M.: Web search strategies (2000) 0.01
    0.01211905 = product of:
      0.018178575 = sum of:
        0.00593598 = weight(_text_:a in 1188) [ClassicSimilarity], result of:
          0.00593598 = score(doc=1188,freq=10.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.11394546 = fieldWeight in 1188, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.03125 = fieldNorm(doc=1188)
        0.012242594 = product of:
          0.024485188 = sum of:
            0.024485188 = weight(_text_:22 in 1188) [ClassicSimilarity], result of:
              0.024485188 = score(doc=1188,freq=2.0), product of:
                0.15821345 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045180224 = queryNorm
                0.15476047 = fieldWeight in 1188, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1188)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Surfing the World Wide Web used to be cool, dude, real cool. But things have gotten hot - so hot that finding something useful an the Web is no longer cool. It is suffocating Web searchers in the smoke and debris of mountain-sized lists of hits, decisions about which search engines they should use, whether they will get lost in the dizzying maze of a subject directory, use the right syntax for the search engine at hand, enter keywords that are likely to retrieve hits an the topics they have in mind, or enlist a browser that has sufficient functionality to display the most promising hits. When it comes to Web searching, in a few short years we have gone from the cool image of surfing the Web into the frying pan of searching the Web. We can turn down the heat by rethinking what Web searchers are doing and introduce some order into the chaos. Web search strategies that are tool-based-oriented to specific Web searching tools such as search en gines, subject directories, and meta search engines-have been widely promoted, and these strategies are just not working. It is time to dissect what Web searching tools expect from searchers and adjust our search strategies to these new tools. This discussion offers Web searchers help in the form of search strategies that are based an strategies that librarians have been using for a long time to search commercial information retrieval systems like Dialog, NEXIS, Wilsonline, FirstSearch, and Data-Star.
    Date
    22. 9.1997 19:16:05
    Type
    a
  7. Herrera-Viedma, E.; Pasi, G.: Soft approaches to information retrieval and information access on the Web : an introduction to the special topic section (2006) 0.01
    0.0117012635 = product of:
      0.017551895 = sum of:
        0.0053093014 = weight(_text_:a in 5285) [ClassicSimilarity], result of:
          0.0053093014 = score(doc=5285,freq=8.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.10191591 = fieldWeight in 5285, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.03125 = fieldNorm(doc=5285)
        0.012242594 = product of:
          0.024485188 = sum of:
            0.024485188 = weight(_text_:22 in 5285) [ClassicSimilarity], result of:
              0.024485188 = score(doc=5285,freq=2.0), product of:
                0.15821345 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045180224 = queryNorm
                0.15476047 = fieldWeight in 5285, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5285)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The World Wide Web is a popular and interactive medium used to collect, disseminate, and access an increasingly huge amount of information, which constitutes the mainstay of the so-called information and knowledge society. Because of its spectacular growth, related to both Web resources (pages, sites, and services) and number of users, the Web is nowadays the main information repository and provides some automatic systems for locating, accessing, and retrieving information. However, an open and crucial question remains: how to provide fast and effective retrieval of the information relevant to specific users' needs. This is a very hard and complex task, since it is pervaded with subjectivity, vagueness, and uncertainty. The expression soft computing refers to techniques and methodologies that work synergistically with the aim of providing flexible information processing tolerant of imprecision, vagueness, partial truth, and approximation. So, soft computing represents a good candidate to design effective systems for information access and retrieval on the Web. One of the most representative tools of soft computing is fuzzy set theory. This special topic section collects research articles witnessing some recent advances in improving the processes of information access and retrieval on the Web by using soft computing tools, and in particular, by using fuzzy sets and/or integrating them with other soft computing tools. In this introductory article, we first review the problem of Web retrieval and the concept of soft computing technology. We then briefly introduce the articles in this section and conclude by highlighting some future research directions that could benefit from the use of soft computing technologies.
    Date
    22. 7.2006 16:59:33
    Type
    a
  8. Vise, D.A.; Malseed, M.: ¬The Google story (2005) 0.01
    0.011521464 = product of:
      0.017282195 = sum of:
        0.0065699257 = weight(_text_:a in 5937) [ClassicSimilarity], result of:
          0.0065699257 = score(doc=5937,freq=16.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.12611452 = fieldWeight in 5937, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.02734375 = fieldNorm(doc=5937)
        0.01071227 = product of:
          0.02142454 = sum of:
            0.02142454 = weight(_text_:22 in 5937) [ClassicSimilarity], result of:
              0.02142454 = score(doc=5937,freq=2.0), product of:
                0.15821345 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045180224 = queryNorm
                0.1354154 = fieldWeight in 5937, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=5937)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Social phenomena happen, and the historians follow. So it goes with Google, the latest star shooting through the universe of trend-setting businesses. This company has even entered our popular lexicon: as many note, "Google" has moved beyond noun to verb, becoming an action which most tech-savvy citizens at the turn of the twenty-first century recognize and in fact do, on a daily basis. It's this wide societal impact that fascinated authors David Vise and Mark Malseed, who came to the book with well-established reputations in investigative reporting. Vise authored the bestselling The Bureau and the Mole, and Malseed contributed significantly to two Bob Woodward books, Bush at War and Plan of Attack. The kind of voluminous research and behind-the-scenes insight in which both writers specialize, and on which their earlier books rested, comes through in The Google Story. The strength of the book comes from its command of many small details, and its focus on the human side of the Google story, as opposed to the merely academic one. Some may prefer a dryer, more analytic approach to Google's impact on the Internet, like The Search or books that tilt more heavily towards bits and bytes on the spectrum between technology and business, like The Singularity is Near. Those wanting to understand the motivations and personal growth of founders Larry Page and Sergey Brin and CEO Eric Schmidt, however, will enjoy this book. Vise and Malseed interviewed over 150 people, including numerous Google employees, Wall Street analysts, Stanford professors, venture capitalists, even Larry Page's Cub Scout leader, and their comprehensiveness shows. As the narrative unfolds, readers learn how Google grew out of the intellectually fertile and not particularly directed friendship between Page and Brin; how the founders attempted to peddle early versions of their search technology to different Silicon Valley firms for $1 million; how Larry and Sergey celebrated their first investor's check with breakfast at Burger King; how the pair initially housed their company in a Palo Alto office, then eventually moved to a futuristic campus dubbed the "Googleplex"; how the company found its financial footing through keyword-targeted Web ads; how various products like Google News, Froogle, and others were cooked up by an inventive staff; how Brin and Page proved their mettle as tough businessmen through negotiations with AOL Europe and their controversial IPO process, among other instances; and how the company's vision for itself continues to grow, such as geographic expansion to China and cooperation with Craig Venter on the Human Genome Project. Like the company it profiles, The Google Story is a bit of a wild ride, and fun, too. Its first appendix lists 23 "tips" which readers can use to get more utility out of Google. The second contains the intelligence test which Google Research offers to prospective job applicants, and shows the sometimes zany methods of this most unusual business. Through it all, Vise and Malseed synthesize a variety of fascinating anecdotes and speculation about Google, and readers seeking a first draft of the history of the company will enjoy an easy read.
    Date
    3. 5.1997 8:44:22
  9. bbu/c't: Ask Jeeves mit verbesserten Suchfunktionen (2005) 0.01
    0.0061459886 = product of:
      0.018437965 = sum of:
        0.018437965 = product of:
          0.03687593 = sum of:
            0.03687593 = weight(_text_:de in 3453) [ClassicSimilarity], result of:
              0.03687593 = score(doc=3453,freq=2.0), product of:
                0.19416152 = queryWeight, product of:
                  4.297489 = idf(docFreq=1634, maxDocs=44218)
                  0.045180224 = queryNorm
                0.18992399 = fieldWeight in 3453, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.297489 = idf(docFreq=1634, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3453)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Mit nicht völlig neuen, aber überarbeiteten Suchfunktionen erweitert das zum Firmenimperium des US-Medienzaren Barry Diller gehörende Unternehmen Ask Jeeves das Leistungsspektrum seiner Suchmaschine. Mit der Ergebnisverfeinerungsfunktion Focus erhält der Suchende auf der rechten oberen Bildschirmseite eine Liste, die das Thema seiner Suche thematisch aufgliedern soll. Eine zweite Neuerung verspricht präzise Antworten auf als Fragen formulierte Sucheinträge. So ergibt der Eintrag "Lady Diana" zum Beispiel eine Liste mit den Items Princess Di, Princess Dianas Life, Princess Diana's Wedding. Interessant dabei ist, dass diese Liste nicht einfach aus einem monolithischen Block von Schlüsselwörtern besteht, sondern in drei Kategorien aufgeteilt ist: "Narrow Your Search", "Expand Your Search" und "Related Names". Waren die eben genannten Beispiele aus der ersten Kategorie, finden sich unter Expand Your Search Einträge wie Royal Family, Princess Di Ring, Princess Di Prince Charles History oder Prince William Harry, allerdings auch Who Is Louis De Funes? "Related Names" verweist auf Einträge wie Diana Spencer, Prince Harry oder Imran Khan. Die Suchfunktion soll also die thematische Verfeinerung oder Ausweitung gleichermaßen wie die Fortsetzung der Suche mit einem verwandten Thema ermöglichen. Auf die Frage "who invented the telephone" erhält der Suchende als ersten Eintrag die Antwort "The telephone was invented by Alexander Graham Bell" mit dem roten Vermerk "Web Answer'. Bemerkenswert ist hier, dass auf eine Frage nicht nur eine passende Webseite mit der Antwort angezeigt wird, sondern die ausformulierte Antwort direkt aus der vorgeschlagenen Webseite zitiert wird. Die Frage "who is the mother of Albert Einstein" gibt immerhin einen Eintrag unter "Narrow Your Search" mit "Albert Einstein Family tree". Ask Jeeves wird wohl noch eine weitere Neuerung bevorstehen: Auf einer Pressekonferenz in San Francisco bemerkte Chief Executive Barry Diller, dass das Unternehmen über eine Namensänderung von Ask Jeeves nachdenke. Wahrscheinlich werde auf eines der beiden Worte verzichtet werden. Mit dem Sucheintrag "How will Ask Jeeves be called in the future" erhält man bislang jedoch noch keine "Web Answer". (26.05.2005 15:30)
  10. Hock, R.: ¬A new era of search engines : not just Web pages anymore (2002) 0.00
    0.0043799505 = product of:
      0.013139851 = sum of:
        0.013139851 = weight(_text_:a in 7688) [ClassicSimilarity], result of:
          0.013139851 = score(doc=7688,freq=4.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.25222903 = fieldWeight in 7688, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=7688)
      0.33333334 = coord(1/3)
    
    Type
    a
  11. Sugiura, A.; Etzioni, O.: Query routing for Web search engines : architecture and experiments (2000) 0.00
    0.0043799505 = product of:
      0.013139851 = sum of:
        0.013139851 = weight(_text_:a in 5009) [ClassicSimilarity], result of:
          0.013139851 = score(doc=5009,freq=4.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.25222903 = fieldWeight in 5009, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=5009)
      0.33333334 = coord(1/3)
    
    Type
    a
  12. Hewett, S.: MathGate - a gateway to Internet resources for mathematicians (2000) 0.00
    0.0043799505 = product of:
      0.013139851 = sum of:
        0.013139851 = weight(_text_:a in 4877) [ClassicSimilarity], result of:
          0.013139851 = score(doc=4877,freq=4.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.25222903 = fieldWeight in 4877, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=4877)
      0.33333334 = coord(1/3)
    
    Type
    a
  13. Sauperl, A.; Say, J.D.: When 'surfing' the Web isn't good enough : providing access to electronic resources (2001) 0.00
    0.0043799505 = product of:
      0.013139851 = sum of:
        0.013139851 = weight(_text_:a in 6951) [ClassicSimilarity], result of:
          0.013139851 = score(doc=6951,freq=4.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.25222903 = fieldWeight in 6951, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=6951)
      0.33333334 = coord(1/3)
    
    Type
    a
  14. Zhang, J.; Dimitroff, A.: Internet search engines' response to Metadata Dublin Core implementation (2005) 0.00
    0.0043799505 = product of:
      0.013139851 = sum of:
        0.013139851 = weight(_text_:a in 4652) [ClassicSimilarity], result of:
          0.013139851 = score(doc=4652,freq=4.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.25222903 = fieldWeight in 4652, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.109375 = fieldNorm(doc=4652)
      0.33333334 = coord(1/3)
    
    Type
    a
  15. Bar-Ilan, J.: Methods for measuring search engine performance over time (2002) 0.00
    0.0043350267 = product of:
      0.01300508 = sum of:
        0.01300508 = weight(_text_:a in 305) [ClassicSimilarity], result of:
          0.01300508 = score(doc=305,freq=12.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.24964198 = fieldWeight in 305, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0625 = fieldNorm(doc=305)
      0.33333334 = coord(1/3)
    
    Abstract
    This study introduces methods for evaluating search engine performance over a time period. Several measures are defined, which as a whole describe search engine functionality over time. The necessary setup for such studies is described, and the use of these measures is illustrated through a specific example. The set of measures introduced here may serve as a guideline for the search engines for testing and improving their functionality. We recommend setting up a standard suite of measures for evaluating search engine performance.
    Type
    a
  16. Watters, C.; Amoudi, A.: Geosearcher : location-based ranking of search engine results (2003) 0.00
    0.0042839246 = product of:
      0.012851773 = sum of:
        0.012851773 = weight(_text_:a in 5152) [ClassicSimilarity], result of:
          0.012851773 = score(doc=5152,freq=30.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.24669915 = fieldWeight in 5152, product of:
              5.477226 = tf(freq=30.0), with freq of:
                30.0 = termFreq=30.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5152)
      0.33333334 = coord(1/3)
    
    Abstract
    Waters and Amoudi describe GeoSearcher, a prototype ranking program that arranges search engine results along a geo-spatial dimension without the provision of geo-spatial meta-tags or the use of geo-spatial feature extraction. GeoSearcher uses URL analysis, IptoLL, Whois, and the Getty Thesaurus of Geographic Names to determine site location. It accepts the first 200 sites returned by a search engine, identifies the coordinates, calculates their distance from a reference point and ranks in ascending order by this value. For any retrieved site the system checks if it has already been located in the current session, then sends the domain name to Whois to generate a return of a two letter country code and an area code. With no success the name is stripped one level and resent. If this fails the top level domain is tested for being a country code. Any remaining unmatched names go to IptoLL. Distance is calculated using the center point of the geographic area and a provided reference location. A test run on a set of 100 URLs from a search was successful in locating 90 sites. Eighty three pages could be manually found and 68 had sufficient information to verify location determination. Of these 65 ( 95%) had been assigned reasonably correct geographic locations. A random set of URLs used instead of a search result, yielded 80% success.
    Type
    a
  17. Mowshowitz, A.; Kawaguchi, A.: Assessing bias in search engines (2002) 0.00
    0.0041973717 = product of:
      0.0125921145 = sum of:
        0.0125921145 = weight(_text_:a in 2574) [ClassicSimilarity], result of:
          0.0125921145 = score(doc=2574,freq=20.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.24171482 = fieldWeight in 2574, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.046875 = fieldNorm(doc=2574)
      0.33333334 = coord(1/3)
    
    Abstract
    This paper deals with the measurement of bias in search engines on the World Wide Web. Bias is taken to mean the balance and representativeness of items in a collection retrieved from a database for a set of queries. This calls for assessing the degree to which the distribution of items in a collection deviates from the ideal. Ascertaining this ideal poses problems similar to those associated with determining relevance in the measurement of recall and precision. Instead of enlisting subject experts or users to determine such an ideal, a family of comparable search engines is used to approximate it for a set of queries. The distribution is obtained by computing the frequencies of occurrence of the uniform resource locators (URLs) in the collection retrieved by several search engines for the given queries. Bias is assessed by measuring the deviation from the ideal of the distribution produced by a particular search engine.
    Type
    a
  18. Zhang, J.; Dimitroff, A.: ¬The impact of metadata implementation on webpage visibility in search engine results : part II (2005) 0.00
    0.0040970687 = product of:
      0.012291206 = sum of:
        0.012291206 = weight(_text_:a in 1027) [ClassicSimilarity], result of:
          0.012291206 = score(doc=1027,freq=14.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.23593865 = fieldWeight in 1027, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1027)
      0.33333334 = coord(1/3)
    
    Abstract
    This paper discusses the impact of metadata implementation in a webpage on its visibility performance in a search engine results list. Influential internal and external factors of metadata implementation were identified. How these factors affect webpage visibility in a search engine results list was examined in an experimental study. Findings suggest that metadata is a good mechanism to improve webpage visibility, the metadata subject field plays a more important role than any other metadata field and keywords extracted from the webpage itself, particularly title or full-text, are most effective. To maximize the effects, these keywords should come from both title and full-text.
    Type
    a
  19. Zhang, J.; Dimitroff, A.: ¬The impact of metadata implementation on webpage visibility in search engine results : part II (2005) 0.00
    0.0040970687 = product of:
      0.012291206 = sum of:
        0.012291206 = weight(_text_:a in 1033) [ClassicSimilarity], result of:
          0.012291206 = score(doc=1033,freq=14.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.23593865 = fieldWeight in 1033, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1033)
      0.33333334 = coord(1/3)
    
    Abstract
    This paper discusses the impact of metadata implementation in a webpage on its visibility performance in a search engine results list. Influential internal and external factors of metadata implementation were identified. How these factors affect webpage visibility in a search engine results list was examined in an experimental study. Findings suggest that metadata is a good mechanism to improve webpage visibility, the metadata subject field plays a more important role than any other metadata field and keywords extracted from the webpage itself, particularly title or full-text, are most effective. To maximize the effects, these keywords should come from both title and full-text.
    Type
    a
  20. Notess, G.R.: Custom search engines : tools and tips (2008) 0.00
    0.0040970687 = product of:
      0.012291206 = sum of:
        0.012291206 = weight(_text_:a in 2145) [ClassicSimilarity], result of:
          0.012291206 = score(doc=2145,freq=14.0), product of:
            0.05209492 = queryWeight, product of:
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.045180224 = queryNorm
            0.23593865 = fieldWeight in 2145, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.153047 = idf(docFreq=37942, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2145)
      0.33333334 = coord(1/3)
    
    Abstract
    The basic steps to build one are fairly simple: * Sign up * Pick a search engine name * Choose a list of sites * Add the sites * Publish That quickly, a search engine can be created to search a specific portion of the web, such as local government sites, childcare resources, or historical archives. It is easy to create a simple customized vertical search engine as well as support much more advanced capabilities (see the Google AJAX search API article). Try these tools and tips and build a customized search engine or two for your own users to help them find more targeted and relevant web information.
    Type
    a

Types

  • a 251
  • el 25
  • m 12
  • r 2
  • s 1
  • x 1
  • More… Less…