Search (725 results, page 1 of 37)

  • × theme_ss:"Suchmaschinen"
  1. Li, L.; Shang, Y.; Zhang, W.: Improvement of HITS-based algorithms on Web documents 0.33
    0.32620016 = product of:
      0.6524003 = sum of:
        0.04544258 = product of:
          0.13632774 = sum of:
            0.13632774 = weight(_text_:3a in 2514) [ClassicSimilarity], result of:
              0.13632774 = score(doc=2514,freq=2.0), product of:
                0.24256827 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.028611459 = queryNorm
                0.56201804 = fieldWeight in 2514, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2514)
          0.33333334 = coord(1/3)
        0.19279654 = weight(_text_:2f in 2514) [ClassicSimilarity], result of:
          0.19279654 = score(doc=2514,freq=4.0), product of:
            0.24256827 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.028611459 = queryNorm
            0.7948135 = fieldWeight in 2514, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=2514)
        0.19279654 = weight(_text_:2f in 2514) [ClassicSimilarity], result of:
          0.19279654 = score(doc=2514,freq=4.0), product of:
            0.24256827 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.028611459 = queryNorm
            0.7948135 = fieldWeight in 2514, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=2514)
        0.028568096 = weight(_text_:web in 2514) [ClassicSimilarity], result of:
          0.028568096 = score(doc=2514,freq=4.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.3059541 = fieldWeight in 2514, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=2514)
        0.19279654 = weight(_text_:2f in 2514) [ClassicSimilarity], result of:
          0.19279654 = score(doc=2514,freq=4.0), product of:
            0.24256827 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.028611459 = queryNorm
            0.7948135 = fieldWeight in 2514, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=2514)
      0.5 = coord(5/10)
    
    Content
    Vgl.: http%3A%2F%2Fdelab.csd.auth.gr%2F~dimitris%2Fcourses%2Fir_spring06%2Fpage_rank_computing%2Fp527-li.pdf. Vgl. auch: http://www2002.org/CDROM/refereed/643/.
    Source
    WWW '02: Proceedings of the 11th International Conference on World Wide Web, May 7-11, 2002, Honolulu, Hawaii, USA
  2. ¬Der große, exklusive TOMORROW-Text : Die beste Suchmaschine der Welt ... und der beste Web-Katalog ... und der beste Metasucher (2000) 0.04
    0.04440908 = product of:
      0.14803027 = sum of:
        0.023567477 = weight(_text_:web in 1522) [ClassicSimilarity], result of:
          0.023567477 = score(doc=1522,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.25239927 = fieldWeight in 1522, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1522)
        0.11533561 = weight(_text_:schutz in 1522) [ClassicSimilarity], result of:
          0.11533561 = score(doc=1522,freq=2.0), product of:
            0.20656188 = queryWeight, product of:
              7.2195506 = idf(docFreq=87, maxDocs=44218)
              0.028611459 = queryNorm
            0.55835867 = fieldWeight in 1522, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.2195506 = idf(docFreq=87, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1522)
        0.009127174 = product of:
          0.027381519 = sum of:
            0.027381519 = weight(_text_:29 in 1522) [ClassicSimilarity], result of:
              0.027381519 = score(doc=1522,freq=2.0), product of:
                0.10064617 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.028611459 = queryNorm
                0.27205724 = fieldWeight in 1522, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1522)
          0.33333334 = coord(1/3)
      0.3 = coord(3/10)
    
    Content
    Darunter Einzel-Beiträge zu: Acoon, Yahoo, MetaGer; Interviews mit den Suchmaschinen-Bossen zu: Wer ist der Lieblingskonkurrent?; So arbeitet eine Suchmaschine; KARZAUNINKAT, S.: So einfach finden sie, was Sie gerade wollen; 20 Fragen: Welcher Suchmaschinen-Typ sind Sie?; KARZAUNINKAT, S.: Kontrolle ist der beste Schutz; BETZ, S.: Darum suchen Sie kostenlos; GLASER, S.: Zwischen Quatsch und Quantenphysik; Suchmaschinen für Spezialfragen
    Date
    29. 4.2000 18:10:50
  3. Jansen, B.J.; Spink, A.; Pedersen, J.: ¬A temporal comparison of AItaVista Web searching (2005) 0.04
    0.04056076 = product of:
      0.13520253 = sum of:
        0.0494814 = weight(_text_:web in 3454) [ClassicSimilarity], result of:
          0.0494814 = score(doc=3454,freq=12.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.5299281 = fieldWeight in 3454, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=3454)
        0.07789783 = weight(_text_:log in 3454) [ClassicSimilarity], result of:
          0.07789783 = score(doc=3454,freq=2.0), product of:
            0.18335998 = queryWeight, product of:
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.028611459 = queryNorm
            0.42483553 = fieldWeight in 3454, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.046875 = fieldNorm(doc=3454)
        0.007823291 = product of:
          0.023469873 = sum of:
            0.023469873 = weight(_text_:29 in 3454) [ClassicSimilarity], result of:
              0.023469873 = score(doc=3454,freq=2.0), product of:
                0.10064617 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.028611459 = queryNorm
                0.23319192 = fieldWeight in 3454, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3454)
          0.33333334 = coord(1/3)
      0.3 = coord(3/10)
    
    Abstract
    Major Web search engines, such as AItaVista, are essential tools in the quest to locate online information. This article reports research that used transaction log analysis to examine the characteristics and changes in AItaVista Web searching that occurred from 1998 to 2002. The research questions we examined are (1) What are the changes in AItaVista Web searching from 1998 to 2002? (2) What are the current characteristics of AItaVista searching, including the duration and frequency of search sessions? (3) What changes in the information needs of AItaVista users occurred between 1998 and 2002? The results of our research show (1) a move toward more interactivity with increases in session and query length, (2) with 70% of session durations at 5 minutes or less, the frequency of interaction is increasing, but it is happening very quickly, and (3) a broadening range of Web searchers' information needs, with the most frequent terms accounting for less than 1% of total term usage. We discuss the implications of these findings for the development of Web search engines.
    Date
    3. 6.2005 19:29:59
  4. Kruschwitz, U.; Lungley, D.; Albakour, M-D.; Song, D.: Deriving query suggestions for site search (2013) 0.04
    0.036563005 = product of:
      0.18281503 = sum of:
        0.023806747 = weight(_text_:web in 1085) [ClassicSimilarity], result of:
          0.023806747 = score(doc=1085,freq=4.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.25496176 = fieldWeight in 1085, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1085)
        0.15900828 = weight(_text_:log in 1085) [ClassicSimilarity], result of:
          0.15900828 = score(doc=1085,freq=12.0), product of:
            0.18335998 = queryWeight, product of:
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.028611459 = queryNorm
            0.86719185 = fieldWeight in 1085, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1085)
      0.2 = coord(2/10)
    
    Abstract
    Modern search engines have been moving away from simplistic interfaces that aimed at satisfying a user's need with a single-shot query. Interactive features are now integral parts of web search engines. However, generating good query modification suggestions remains a challenging issue. Query log analysis is one of the major strands of work in this direction. Although much research has been performed on query logs collected on the web as a whole, query log analysis to enhance search on smaller and more focused collections has attracted less attention, despite its increasing practical importance. In this article, we report on a systematic study of different query modification methods applied to a substantial query log collected on a local website that already uses an interactive search engine. We conducted experiments in which we asked users to assess the relevance of potential query modification suggestions that have been constructed using a range of log analysis methods and different baseline approaches. The experimental results demonstrate the usefulness of log analysis to extract query modification suggestions. Furthermore, our experiments demonstrate that a more fine-grained approach than grouping search requests into sessions allows for extraction of better refinement terms from query log files.
  5. Beuth, P.: ¬Die Jagd nach Nutzer-Profilen (2009) 0.03
    0.032516427 = product of:
      0.10838809 = sum of:
        0.055082086 = weight(_text_:log in 3033) [ClassicSimilarity], result of:
          0.055082086 = score(doc=3033,freq=4.0), product of:
            0.18335998 = queryWeight, product of:
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.028611459 = queryNorm
            0.30040407 = fieldWeight in 3033, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.0234375 = fieldNorm(doc=3033)
        0.049429547 = weight(_text_:schutz in 3033) [ClassicSimilarity], result of:
          0.049429547 = score(doc=3033,freq=2.0), product of:
            0.20656188 = queryWeight, product of:
              7.2195506 = idf(docFreq=87, maxDocs=44218)
              0.028611459 = queryNorm
            0.23929656 = fieldWeight in 3033, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.2195506 = idf(docFreq=87, maxDocs=44218)
              0.0234375 = fieldNorm(doc=3033)
        0.0038764577 = product of:
          0.011629373 = sum of:
            0.011629373 = weight(_text_:22 in 3033) [ClassicSimilarity], result of:
              0.011629373 = score(doc=3033,freq=2.0), product of:
                0.10019246 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.028611459 = queryNorm
                0.116070345 = fieldWeight in 3033, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=3033)
          0.33333334 = coord(1/3)
      0.3 = coord(3/10)
    
    Content
    Zusammen soll das ein Konglomerat ergeben, dass es technisch und inhaltlich mit Google aufnehmen kann. Aufnehmen heißt erstens, die Allgegenwärtigkeit von Google-Diensten im Netz angreifen zu können und möglichst viele Internetnutzer an sich zu binden. Aufnehmen heißt zweitens, diese Nutzer eben so gut ausspionieren zu können wie Google das tut. Bausteine für die komplette Durchleuchtung ihrer Nutzer existieren schon bei allen drei Unternehmen. Es sind kleine, unauffällige Links auf ihren Startseiten. Die Bausteine heißen iGoogle, My MSN und MyYahoo. Sie sind die Zukunft der Internetnutzung, so wie sie sich Google, Yahoo und Microsoft vorstellen. Hinter allen drei Links stehen personalisierbare Startseiten, mit denen sich der Nutzer mehr oder weniger seine gesamte Online-Präsenz in ein einziges Browserfenster holt, das er für seine Standardaktivitäten dann kaum noch verlassen muss. Suchmaschine, aktuelle Nachrichten, E-Mails, Messenger oder virtuelle Notizzettel - alles individuell kombinierbar, und alles auf einen Blick. Bezahlen muss der Nutzer dafür nicht, Mitmachen kostet nur eine E-Mail-Adresse und ein Passwort. Nach dem Log-in beginnt das Profiling. Google arbeitet sogar schon am übernächsten Schritt. Der Konzern hat jüngst ein eigenes Betriebssystem angekündigt, das weitgehend netzbasiert arbeiten soll. Dann werden im Browserfenster zusätzlich noch Textverarbeitungsprogramme und Media-Player laufen. Briefe und Tabellen speichert der Nutzer dann nicht mehr auf der Festplatte ab, sondern auf Googles Servern.
    So entstehen noch viel mehr Daten, die einem Nutzer zugeordnet werden können. Wie genau diese weiterverarbeitet und an dritte verkauft werden, kann niemand nachvollziehen. Wer sich darauf einlässt, gibt die Hoheit über seine persönlichen Daten ab und damit ein Grundrecht. Aus informationeller Selbstbestimmung wird informationelle Fremdbestimmung. Es ist natürlich das gute Recht eines jeden, dieses Grundrecht in den Wind zu schlagen. Für viele, nicht nur junge Menschen, ist der Austausch von Informationen aller Art selbstverständlich geworden, um Kontakte zu erhalten und soziale und geschäftliche Netze zu knüpfen. Diejenigen, die andere per Internet nicht nur jederzeit über ihren Gemütszustand, sondern sogar über ihren genauen Aufenthaltsort unterrichten, empfinden personalisierte Werbung möglicherweise sogar als angenehm. Wer aber den Gedanken unheimlich findet, dass jede Firma der Welt einen Datensatz erwerben kann, der Auskunft über seine Wünsche, Neigungen, Fantasien, Krankheiten und Beziehungen gibt, der muss den Log-in verweigern, auch wenn alle Welt freudig mitmacht. Der Preis für den Selbst(-Daten)schutz kann durchaus die soziale Ausgrenzung im Netz sein. Der Internetnutzer der Zukunft wird - so lautet der Plan von Google, Yahoo und Microsoft - einer großen Versuchung ausgesetzt sein: Alles aus einer Hand. Oder anders gesagt: Alles in eine Hand."
    Date
    17. 7.1996 9:33:22
  6. Koshman, S.; Spink, A.; Jansen, B.J.: Web searching on the Vivisimo search engine (2006) 0.03
    0.027268365 = product of:
      0.13634183 = sum of:
        0.044538345 = weight(_text_:web in 216) [ClassicSimilarity], result of:
          0.044538345 = score(doc=216,freq=14.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.47698978 = fieldWeight in 216, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=216)
        0.091803476 = weight(_text_:log in 216) [ClassicSimilarity], result of:
          0.091803476 = score(doc=216,freq=4.0), product of:
            0.18335998 = queryWeight, product of:
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.028611459 = queryNorm
            0.5006735 = fieldWeight in 216, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.0390625 = fieldNorm(doc=216)
      0.2 = coord(2/10)
    
    Abstract
    The application of clustering to Web search engine technology is a novel approach that offers structure to the information deluge often faced by Web searchers. Clustering methods have been well studied in research labs; however, real user searching with clustering systems in operational Web environments is not well understood. This article reports on results from a transaction log analysis of Vivisimo.com, which is a Web meta-search engine that dynamically clusters users' search results. A transaction log analysis was conducted on 2-week's worth of data collected from March 28 to April 4 and April 25 to May 2, 2004, representing 100% of site traffic during these periods and 2,029,734 queries overall. The results show that the highest percentage of queries contained two terms. The highest percentage of search sessions contained one query and was less than 1 minute in duration. Almost half of user interactions with clusters consisted of displaying a cluster's result set, and a small percentage of interactions showed cluster tree expansion. Findings show that 11.1% of search sessions were multitasking searches, and there are a broad variety of search topics in multitasking search sessions. Other searching interactions and statistics on repeat users of the search engine are reported. These results provide insights into search characteristics with a cluster-based Web search engine and extend research into Web searching trends.
  7. Wikia Search : Das Anti-Google (2008) 0.03
    0.02646286 = product of:
      0.08820953 = sum of:
        0.016833913 = weight(_text_:web in 369) [ClassicSimilarity], result of:
          0.016833913 = score(doc=369,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.18028519 = fieldWeight in 369, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=369)
        0.06491486 = weight(_text_:log in 369) [ClassicSimilarity], result of:
          0.06491486 = score(doc=369,freq=2.0), product of:
            0.18335998 = queryWeight, product of:
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.028611459 = queryNorm
            0.3540296 = fieldWeight in 369, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.0390625 = fieldNorm(doc=369)
        0.006460763 = product of:
          0.019382289 = sum of:
            0.019382289 = weight(_text_:22 in 369) [ClassicSimilarity], result of:
              0.019382289 = score(doc=369,freq=2.0), product of:
                0.10019246 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.028611459 = queryNorm
                0.19345059 = fieldWeight in 369, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=369)
          0.33333334 = coord(1/3)
      0.3 = coord(3/10)
    
    Content
    "Es ist ein Start mit verschiedenen Schwächen: Der neue Google-Konkurrent "Wikia Search" verschreckt den User in seiner aktuellen Vor-abversion mit fehlerhaften Links, fehlenden Ergebnissen oder auch fragwürdigen Seitenbewertungen. Trotzdem bietet die neue Suchmaschine des Wikipedia-Gründers Jimmy Wales schon jetzt einen unschätzbaren Vorteil gegenüber dein Branchenprimus Google: Wikia Search legt den Nutzern offen, wie die Suchergebnisse zustande kommen. Sie setzt nämlich OpenSource-Programme ein - etwa den Webcrawler Grub. Dessen Algorithmen liegen für jedermann offen einsehbar im Web. Wikia Search fordert gar zum aktiven Mitmachen auf: Die Surfer können angezeigte Links bewerten Lind so die Rangfolge der Ergebnisse beeinflussen. Die Suchmaschine soll die Ergebnisse mit der Zeit besser nach Relevanz und Qualität sortieren, verspricht das Entwicklerteam. Derzeit speichert die Seite Bewertungen durch User jedoch noch nicht ab. Zusätzlich zeigt Wikia Search, passend zu den Suchbegriffen, jeweils einen kurzen Übersichtsartikel an, der allgemeine Informationen enthält. Sollte so ein Text noch fehlen, kann ganz einfach per Mausklick ein neues Dokument angelegt werden. Praktisch dabei: Jeder Nutzer hat die Möglichkeit, den Inhalt ohne Anmeldung zu bearbeiten und zu erweitern. Eine Alternative zur normalen Suche soll die »Whitelist« sein. Darin können User eine Art Webseiten-Kata-log anlegen, der für bessere Suchergebnisse sorgen soll. Die Idee einer offenen und nutzerorientierten Suchmaschine klingt vielversprechend - eine ernsthafte Konkurrenz für etablierte Suchmaschinen wie Google kann Wikia Search allerdings noch nicht sein. Denn die neue Websuche ist abhängig von der Community und ihrer Mitarbeit an dem Projekt. Dass so ein Ansatz funktionieren kann, hat die Online-Enzyklopädie Wikipedia immerhin gezeigt."
    Source
    Chip. 2008, H.3, S.22
  8. Web search engine research (2012) 0.03
    0.025475848 = product of:
      0.12737924 = sum of:
        0.0494814 = weight(_text_:web in 478) [ClassicSimilarity], result of:
          0.0494814 = score(doc=478,freq=12.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.5299281 = fieldWeight in 478, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=478)
        0.07789783 = weight(_text_:log in 478) [ClassicSimilarity], result of:
          0.07789783 = score(doc=478,freq=2.0), product of:
            0.18335998 = queryWeight, product of:
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.028611459 = queryNorm
            0.42483553 = fieldWeight in 478, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.046875 = fieldNorm(doc=478)
      0.2 = coord(2/10)
    
    Abstract
    "Web Search Engine Research", edited by Dirk Lewandowski, provides an understanding of Web search engines from the unique perspective of Library and Information Science. The book explores a range of topics including retrieval effectiveness, user satisfaction, the evaluation of search interfaces, the impact of search on society, reliability of search results, query log analysis, user guidance in the search process, and the influence of search engine optimization (SEO) on results quality. While research in computer science has mainly focused on technical aspects of search engines, LIS research is centred on users' behaviour when using search engines and how this interaction can be evaluated. LIS research provides a unique perspective in intermediating between the technical aspects, user aspects and their impact on their role in knowledge acquisition. This book is directly relevant to researchers and practitioners in library and information science, computer science, including Web researchers.
    LCSH
    Web search engines
    Subject
    Web search engines
  9. Chau, M.; Fang, X.; Sheng, O.R.U.: Analysis of the query logs of a Web site search engine (2005) 0.02
    0.023629673 = product of:
      0.118148364 = sum of:
        0.053233504 = weight(_text_:web in 4573) [ClassicSimilarity], result of:
          0.053233504 = score(doc=4573,freq=20.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.5701118 = fieldWeight in 4573, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4573)
        0.06491486 = weight(_text_:log in 4573) [ClassicSimilarity], result of:
          0.06491486 = score(doc=4573,freq=2.0), product of:
            0.18335998 = queryWeight, product of:
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.028611459 = queryNorm
            0.3540296 = fieldWeight in 4573, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4573)
      0.2 = coord(2/10)
    
    Abstract
    A large number of studies have investigated the transaction log of general-purpose search engines such as Excite and AItaVista, but few studies have reported an the analysis of search logs for search engines that are limited to particular Web sites, namely, Web site search engines. In this article, we report our research an analyzing the search logs of the search engine of the Utah state government Web site. Our results show that some statistics, such as the number of search terms per query, of Web users are the same for general-purpose search engines and Web site search engines, but others, such as the search topics and the terms used, are considerably different. Possible reasons for the differences include the focused domain of Web site search engines and users' different information needs. The findings are useful for Web site developers to improve the performance of their services provided an the Web and for researchers to conduct further research in this area. The analysis also can be applied in e-government research by investigating how information should be delivered to users in government Web sites.
  10. Shi, X.; Yang, C.C.: Mining related queries from Web search engine query logs using an improved association rule mining model (2007) 0.02
    0.020511327 = product of:
      0.10255663 = sum of:
        0.037641775 = weight(_text_:web in 597) [ClassicSimilarity], result of:
          0.037641775 = score(doc=597,freq=10.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.40312994 = fieldWeight in 597, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=597)
        0.06491486 = weight(_text_:log in 597) [ClassicSimilarity], result of:
          0.06491486 = score(doc=597,freq=2.0), product of:
            0.18335998 = queryWeight, product of:
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.028611459 = queryNorm
            0.3540296 = fieldWeight in 597, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.0390625 = fieldNorm(doc=597)
      0.2 = coord(2/10)
    
    Abstract
    With the overwhelming volume of information, the task of finding relevant information on a given topic on the Web is becoming increasingly difficult. Web search engines hence become one of the most popular solutions available on the Web. However, it has never been easy for novice users to organize and represent their information needs using simple queries. Users have to keep modifying their input queries until they get expected results. Therefore, it is often desirable for search engines to give suggestions on related queries to users. Besides, by identifying those related queries, search engines can potentially perform optimizations on their systems, such as query expansion and file indexing. In this work we propose a method that suggests a list of related queries given an initial input query. The related queries are based in the query log of previously submitted queries by human users, which can be identified using an enhanced model of association rules. Users can utilize the suggested related queries to tune or redirect the search process. Our method not only discovers the related queries, but also ranks them according to the degree of their relatedness. Unlike many other rival techniques, it also performs reasonably well on less frequent input queries.
    Footnote
    Beitrag eines Themenschwerpunktes "Mining Web resources for enhancing information retrieval"
  11. Option für Metager als Standardsuchmaschine, Suchmaschine nach dem Peer-to-Peer-Prinzip (2021) 0.02
    0.0198628 = product of:
      0.099314004 = sum of:
        0.033407938 = weight(_text_:kommunikation in 431) [ClassicSimilarity], result of:
          0.033407938 = score(doc=431,freq=2.0), product of:
            0.14706601 = queryWeight, product of:
              5.140109 = idf(docFreq=703, maxDocs=44218)
              0.028611459 = queryNorm
            0.22716287 = fieldWeight in 431, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.140109 = idf(docFreq=703, maxDocs=44218)
              0.03125 = fieldNorm(doc=431)
        0.06590606 = weight(_text_:schutz in 431) [ClassicSimilarity], result of:
          0.06590606 = score(doc=431,freq=2.0), product of:
            0.20656188 = queryWeight, product of:
              7.2195506 = idf(docFreq=87, maxDocs=44218)
              0.028611459 = queryNorm
            0.31906208 = fieldWeight in 431, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.2195506 = idf(docFreq=87, maxDocs=44218)
              0.03125 = fieldNorm(doc=431)
      0.2 = coord(2/10)
    
    Content
    Auch auf dem Volla-Phone ist es bald möglich, MetaGer als Standardsuchmaschine zu wählen. Das Volla Phone ist ein Produkt von "Hallo Welt Systeme UG" in Remscheid. Die Entwickler des Smartphones verfolgen den Ansatz, möglichst wenig von der Aufmerksamkeit des Nutzers zu beanspruchen. Technik soll nicht ablenken und sich in der Vordergrund spielen, sondern als bloßes Werkzeug im Hintergrund bleiben. Durch Möglichkeiten wie detaillierter Datenschutzeinstellungen, logfreiem VPN, quelloffener Apps aus einem alternativen App Store wird zudem Schutz der Privatsphäre ermöglicht - ganz ohne Google-Dienste. Durch die Partnerschaft mit MetaGer können die Nutzer von Volla-Phone auch im Bereich Suchmaschine Privatsphärenschutz realisieren. Mehr unter: https://suma-ev.de/mit-metager-auf-dem-volla-phone-suchen/
    YaCy: Suchmaschine nach dem Peer-to-Peer-Prinzip. YaCy ist eine dezentrale, freie Suchmaschine. Die Besonderheit: die freie Suchmaschine läuft nicht auf zentralen Servern eines einzelnen Betreibers, sondern funktioniert nach dem Peer-to-Peer (P2P) Prinzip. Dieses basiert darauf, dass die YaCy-Nutzer aufgerufene Webseiten auf ihrem Computer lokal indexieren. Jeder Nutzer "ercrawlt" sich damit einen kleinen Index, den er durch Kommunikation mit anderen YaCy-Peers teilen kann. Das Programm sorgt dafür, dass durch die kleinen dezentralen Crawler einzelner Nutzer schließlich ein globaler Gesamtindex entsteht. Je mehr Nutzer Teil dieser dezentralen Suche sind, desto größer wird der gemeinsame Index, auf den der einzelne Nutzer dann Zugriff haben kann. Seit kurzem befindet sich YaCy im Verbund unserer abgefragten Suchmaschinen. Wir sind somit auch Teil des Indexes der Suchmaschine.
  12. Song, R.; Luo, Z.; Nie, J.-Y.; Yu, Y.; Hon, H.-W.: Identification of ambiguous queries in web search (2009) 0.02
    0.019619705 = product of:
      0.098098524 = sum of:
        0.020200694 = weight(_text_:web in 2441) [ClassicSimilarity], result of:
          0.020200694 = score(doc=2441,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.21634221 = fieldWeight in 2441, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=2441)
        0.07789783 = weight(_text_:log in 2441) [ClassicSimilarity], result of:
          0.07789783 = score(doc=2441,freq=2.0), product of:
            0.18335998 = queryWeight, product of:
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.028611459 = queryNorm
            0.42483553 = fieldWeight in 2441, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.046875 = fieldNorm(doc=2441)
      0.2 = coord(2/10)
    
    Abstract
    It is widely believed that many queries submitted to search engines are inherently ambiguous (e.g., java and apple). However, few studies have tried to classify queries based on ambiguity and to answer "what the proportion of ambiguous queries is". This paper deals with these issues. First, we clarify the definition of ambiguous queries by constructing the taxonomy of queries from being ambiguous to specific. Second, we ask human annotators to manually classify queries. From manually labeled results, we observe that query ambiguity is to some extent predictable. Third, we propose a supervised learning approach to automatically identify ambiguous queries. Experimental results show that we can correctly identify 87% of labeled queries with the approach. Finally, by using our approach, we estimate that about 16% of queries in a real search log are ambiguous.
  13. Rowlands, I.; Nicholas, D.; Williams, P.; Huntington, P.; Fieldhouse, M.; Gunter, B.; Withey, R.; Jamali, H.R.; Dobrowolski, T.; Tenopir, C.: ¬The Google generation : the information behaviour of the researcher of the future (2008) 0.02
    0.018814411 = product of:
      0.09407205 = sum of:
        0.029157192 = weight(_text_:web in 2017) [ClassicSimilarity], result of:
          0.029157192 = score(doc=2017,freq=6.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.3122631 = fieldWeight in 2017, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2017)
        0.06491486 = weight(_text_:log in 2017) [ClassicSimilarity], result of:
          0.06491486 = score(doc=2017,freq=2.0), product of:
            0.18335998 = queryWeight, product of:
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.028611459 = queryNorm
            0.3540296 = fieldWeight in 2017, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2017)
      0.2 = coord(2/10)
    
    Abstract
    Purpose - This article is an edited version of a report commissioned by the British Library and JISC to identify how the specialist researchers of the future (those born after 1993) are likely to access and interact with digital resources in five to ten years' time. The purpose is to investigate the impact of digital transition on the information behaviour of the Google Generation and to guide library and information services to anticipate and react to any new or emerging behaviours in the most effective way. Design/methodology/approach - The study was virtually longitudinal and is based on a number of extensive reviews of related literature, survey data mining and a deep log analysis of a British Library and a JISC web site intended for younger people. Findings - The study shows that much of the impact of ICTs on the young has been overestimated. The study claims that although young people demonstrate an apparent ease and familiarity with computers, they rely heavily on search engines, view rather than read and do not possess the critical and analytical skills to assess the information that they find on the web. Originality/value - The paper reports on a study that overturns the common assumption that the "Google generation" is the most web-literate.
  14. Lewandowski, D.; Drechsler, J.; Mach, S. von: Deriving query intents from web search engine queries (2012) 0.02
    0.017744321 = product of:
      0.0887216 = sum of:
        0.023806747 = weight(_text_:web in 385) [ClassicSimilarity], result of:
          0.023806747 = score(doc=385,freq=4.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.25496176 = fieldWeight in 385, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=385)
        0.06491486 = weight(_text_:log in 385) [ClassicSimilarity], result of:
          0.06491486 = score(doc=385,freq=2.0), product of:
            0.18335998 = queryWeight, product of:
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.028611459 = queryNorm
            0.3540296 = fieldWeight in 385, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.0390625 = fieldNorm(doc=385)
      0.2 = coord(2/10)
    
    Abstract
    The purpose of this article is to test the reliability of query intents derived from queries, either by the user who entered the query or by another juror. We report the findings of three studies. First, we conducted a large-scale classification study (~50,000 queries) using a crowdsourcing approach. Next, we used clickthrough data from a search engine log and validated the judgments given by the jurors from the crowdsourcing study. Finally, we conducted an online survey on a commercial search engine's portal. Because we used the same queries for all three studies, we also were able to compare the results and the effectiveness of the different approaches. We found that neither the crowdsourcing approach, using jurors who classified queries originating from other users, nor the questionnaire approach, using searchers who were asked about their own query that they just entered into a Web search engine, led to satisfying results. This leads us to conclude that there was little understanding of the classification tasks, even though both groups of jurors were given detailed instructions. Although we used manual classification, our research also has important implications for automatic classification. We must question the success of approaches using automatic classification and comparing its performance to a baseline from human jurors.
  15. Sarigil, E.; Sengor Altingovde, I.; Blanco, R.; Barla Cambazoglu, B.; Ozcan, R.; Ulusoy, Ö.: Characterizing, predicting, and handling web search queries that match very few or no results (2018) 0.02
    0.017744321 = product of:
      0.0887216 = sum of:
        0.023806747 = weight(_text_:web in 4039) [ClassicSimilarity], result of:
          0.023806747 = score(doc=4039,freq=4.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.25496176 = fieldWeight in 4039, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4039)
        0.06491486 = weight(_text_:log in 4039) [ClassicSimilarity], result of:
          0.06491486 = score(doc=4039,freq=2.0), product of:
            0.18335998 = queryWeight, product of:
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.028611459 = queryNorm
            0.3540296 = fieldWeight in 4039, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.4086204 = idf(docFreq=197, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4039)
      0.2 = coord(2/10)
    
    Abstract
    A non-negligible fraction of user queries end up with very few or even no matching results in leading commercial web search engines. In this work, we provide a detailed characterization of such queries and show that search engines try to improve such queries by showing the results of related queries. Through a user study, we show that these query suggestions are usually perceived as relevant. Also, through a query log analysis, we show that the users are dissatisfied after submitting a query that match no results at least 88.5% of the time. As a first step towards solving these no-answer queries, we devised a large number of features that can be used to identify such queries and built machine-learning models. These models can be useful for scenarios such as the mobile- or meta-search, where identifying a query that will retrieve no results at the client device (i.e., even before submitting it to the search engine) may yield gains in terms of the bandwidth usage, power consumption, and/or monetary costs. Experiments over query logs indicate that, despite the heavy skew in class sizes, our models achieve good prediction quality, with accuracy (in terms of area under the curve) up to 0.95.
  16. Schulz, W.; Held, T.: ¬Der Index auf dem Index? : Selbstzensur und Zensur bei Suchmaschinen (2007) 0.02
    0.01638655 = product of:
      0.081932746 = sum of:
        0.07086894 = weight(_text_:kommunikation in 374) [ClassicSimilarity], result of:
          0.07086894 = score(doc=374,freq=4.0), product of:
            0.14706601 = queryWeight, product of:
              5.140109 = idf(docFreq=703, maxDocs=44218)
              0.028611459 = queryNorm
            0.48188522 = fieldWeight in 374, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.140109 = idf(docFreq=703, maxDocs=44218)
              0.046875 = fieldNorm(doc=374)
        0.011063806 = product of:
          0.033191416 = sum of:
            0.033191416 = weight(_text_:29 in 374) [ClassicSimilarity], result of:
              0.033191416 = score(doc=374,freq=4.0), product of:
                0.10064617 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.028611459 = queryNorm
                0.3297832 = fieldWeight in 374, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=374)
          0.33333334 = coord(1/3)
      0.2 = coord(2/10)
    
    Abstract
    Suchmaschinen gelten als Gatekeeper in der öffentlichen Kommunikation. Zumindest für bestimmte Typen von Seitenaufrufen stellen Suchmaschinen den mit Abstand häufigsten Weg des Zugangs zu Internet-Inhalten dar. Deshalb können Beeinflussungen des Index von Suchmaschinen oder der Algorithmen, die die Ergebnislisten steuern, als hochsensibler Eingriff in die Internet-basierte Kommunikation angesehen werden. Dies lenkt die Aufmerksamkeit auf die >Policies< der Suchmaschinenanbieter, auch im Hinblick auf externe Anforderungen, die etwa von Nationalstaaten an sie gerichtet werden. Vor allem der Anbieter Google ist in die Kritik geraten, weil er in seinem Angebot in China Seiten aus dem Index löscht, die von der chinesischen Regierung als staatsgefährdend angesehen werden. Google beruft sich darauf, nur die dortigen Gesetze zu befolgen. Auch in Deutschland werden Seiten gefiltert. In der Internetgemeinde ist dann schnell das Wort >Zensur< zu hören. Im Folgenden soll der Frage nachgegangen werden, wann nach deutschem Verständnis von Zensur gesprochen werden kann. Dabei soll deutlich werden, wo Unterschiede in den nationalstaatlichen Politiken, aber auch bei den Kooperationen der Suchmaschinenanbieter mit den Nationalstaaten bestehen.
    Date
    13. 5.2007 10:29:29
  17. Rieder, B.: Demokratisierung der Suche? : von der Kritik zum gesellschaftlich orientierten Design (2009) 0.02
    0.015880339 = product of:
      0.079401694 = sum of:
        0.04175992 = weight(_text_:kommunikation in 24) [ClassicSimilarity], result of:
          0.04175992 = score(doc=24,freq=2.0), product of:
            0.14706601 = queryWeight, product of:
              5.140109 = idf(docFreq=703, maxDocs=44218)
              0.028611459 = queryNorm
            0.28395358 = fieldWeight in 24, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.140109 = idf(docFreq=703, maxDocs=44218)
              0.0390625 = fieldNorm(doc=24)
        0.037641775 = weight(_text_:web in 24) [ClassicSimilarity], result of:
          0.037641775 = score(doc=24,freq=10.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.40312994 = fieldWeight in 24, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=24)
      0.2 = coord(2/10)
    
    Abstract
    Seit die Techniker der Digital Equipment Company 1995 AltaVista einführten, die erste große Suchmaschine für das World Wide Web, hat sich vieles verändert. 2009 ist das Web die zentrale Plattform für alles, was mit Information und Kommunikation zu tun hat: es bietet Raum für eine Vielfalt von Aktivitäten und Vorgängen, die früher über zahlreiche verschiedene Kanäle verteilt waren. Mit 1,5 Milliarden Nutzern, mehr als einer Trillion Seiten und einer Palette von Services, die von der einfachen Darstellung von Text-basierter Information bis zu hoch entwickelten Applikationen und Multimedia-Technologien reichen, ist das Web der Gegenwart ein Informationsgigant und außerdem zentraler Bestandteil einer kapitalistischen Ökonomie, die sich von einem industriellen zu einem kognitiven Produktionsmodus entwickelt. Da das Web kein eigenes Index- oder Katalogsystem mitbringt, liegt es an den Suchmaschinen, die unübersichtliche Struktur des Web den Nutzern zu erschließen. Obwohl Suchmaschinen komplexe Werkzeuge sind, ist ihre Handhabung überraschend einfach: Eine aus einem oder mehreren Wörtern bestehende Suchanfrage führt zu einer geordneten Liste von Seiten, welche die angegebenen Begriffe enthalten. Es ist kein Wunder, dass Suchmaschinen zu den beliebtesten Internet-Diensten gehören.
  18. Jesdanun, A.: Streitbare Suchmaschine : Polar Rose ermöglicht Internet-Recherche mit Gesichtserkennung (2007) 0.02
    0.01587464 = product of:
      0.079373196 = sum of:
        0.013467129 = weight(_text_:web in 547) [ClassicSimilarity], result of:
          0.013467129 = score(doc=547,freq=2.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.14422815 = fieldWeight in 547, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=547)
        0.06590606 = weight(_text_:schutz in 547) [ClassicSimilarity], result of:
          0.06590606 = score(doc=547,freq=2.0), product of:
            0.20656188 = queryWeight, product of:
              7.2195506 = idf(docFreq=87, maxDocs=44218)
              0.028611459 = queryNorm
            0.31906208 = fieldWeight in 547, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.2195506 = idf(docFreq=87, maxDocs=44218)
              0.03125 = fieldNorm(doc=547)
      0.2 = coord(2/10)
    
    Abstract
    Probleme für den Schutz der Persönlichkeitsrechte wirft das Projekt einer schwedischen Firma auf, die eine Internet-Suchmaschine mit Gesichtserkennung entwickelt. Die Technik der Firma Polar Rose scannt öffentlich verfügbare Fotos ein, sortiert sie nach rund 90 verschiedenen Merkmalen und erstellt so eine Datenbank. Die Suchmaschine soll in der Lage sein, ein beliebiges Foto mit diesen Daten abzugleichen, die Identität der gezeigten Person zu ermitteln und eine Liste mit Web-Seiten zu liefern, auf denen diese Person zu sehen ist. Bei Tests. mit 10 000 Fotos habe es in 95 Prozent der Fälle eine zuverlässige Erkennung gegeben, sagt der Vorstandschef von Polar Rose, Nikolaj Nyholm. Allerdings schränkt er ein, dass die Genauigkeit mit wachsender Datenbasis vermutlich geringer wird, weil bei Millionen und vielleicht Milliarden von Personenfotos die Wahrscheinlichkeit zunimmt, dass sich zwei oder mehr Personen sehr ähnlich sehen. Deshalb sollen die Nutzer des geplanten Internet-Dienstes selbst Informationen beisteuern, etwa die Namen von abgebildeten Personen. Polar Rose verfolgt das Konzept, die zahllosen Fotos, die sich etwa bei Flickr oder Myspace finden, besser durchsuchbar zu machen, als bei der herkömmlichen Bildersuche. Auch Personen, die nur im Hintergrund eines Fotos zu sehen sind, sollen auf diese Weise erfasst werden. Was aber ist, wenn Arbeitgeber, Polizei oder misstrauische Partner auf diese Weise die Anwesenheit einer Person an einem bestimmten Ort aufdecken, die eigentlich vertraulich bleiben sollte? "Ich glaube nicht, dass wir da schon alle Antworten haben", räumt Nyholm ein. Der Leiter der Organisation Privacy International, Simon Davies, sieht sich durch Techniken wie die von Polar Rose in seiner Einschätzung bestätigt, dass es Grenzen für die Internet-Suche geben müsse. Andernfalls werde die Suche im Internet in Dimensionen vorstoßen, "die unendlich mächtiger sind, als wir es uns jemals vorstellen konnten". Davies fordert eine Debatte über eine Begrenzung der Internet-Suche und über ein Mitspracherecht von einzelnen Personen bei der Nutzung ihrer Daten. Die Verfügbarkeit von Fotos im Internet sei kein Freibrief für massenhafte Aufbereitung in Datenbanken.
  19. Marchiori, M.: ¬The quest for correct information on the Web : hyper search engines (1997) 0.02
    0.015432911 = product of:
      0.07716455 = sum of:
        0.04082007 = weight(_text_:web in 7453) [ClassicSimilarity], result of:
          0.04082007 = score(doc=7453,freq=6.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.43716836 = fieldWeight in 7453, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=7453)
        0.036344483 = product of:
          0.054516725 = sum of:
            0.027381519 = weight(_text_:29 in 7453) [ClassicSimilarity], result of:
              0.027381519 = score(doc=7453,freq=2.0), product of:
                0.10064617 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.028611459 = queryNorm
                0.27205724 = fieldWeight in 7453, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=7453)
            0.027135205 = weight(_text_:22 in 7453) [ClassicSimilarity], result of:
              0.027135205 = score(doc=7453,freq=2.0), product of:
                0.10019246 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.028611459 = queryNorm
                0.2708308 = fieldWeight in 7453, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=7453)
          0.6666667 = coord(2/3)
      0.2 = coord(2/10)
    
    Abstract
    Presents a novel method to extract from a web object its hyper informative content, in contrast with current search engines, which only deal with the textual information content. This method is not only valuable per se, but it is shown to be able to considerably increase the precision of current search engines. It integrates with existing search engine technology since it can be implemented on top of every search engine, acting as a post-processor, thus automatically transforming a search engine into its corresponding hyper version. Shows how the hyper information can be usefully employed to face the search engines persuasion problem
    Date
    1. 8.1996 22:08:06
    Footnote
    Contribution to a special issue of papers from the 6th International World Wide Web conference, held 7-11 Apr 1997, Santa Clara, California
    Source
    Computer networks and ISDN systems. 29(1997) no.8, S.1225-1235
  20. Carrière, S.J.; Kazman, R.: Webquery : searching and visualising the Web through connectivity (1997) 0.02
    0.015264508 = product of:
      0.07632254 = sum of:
        0.04517013 = weight(_text_:web in 2674) [ClassicSimilarity], result of:
          0.04517013 = score(doc=2674,freq=10.0), product of:
            0.0933738 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.028611459 = queryNorm
            0.48375595 = fieldWeight in 2674, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=2674)
        0.031152412 = product of:
          0.04672862 = sum of:
            0.023469873 = weight(_text_:29 in 2674) [ClassicSimilarity], result of:
              0.023469873 = score(doc=2674,freq=2.0), product of:
                0.10064617 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.028611459 = queryNorm
                0.23319192 = fieldWeight in 2674, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2674)
            0.023258746 = weight(_text_:22 in 2674) [ClassicSimilarity], result of:
              0.023258746 = score(doc=2674,freq=2.0), product of:
                0.10019246 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.028611459 = queryNorm
                0.23214069 = fieldWeight in 2674, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2674)
          0.6666667 = coord(2/3)
      0.2 = coord(2/10)
    
    Abstract
    The WebQuery system offers a powerful new method for searching the Web based on connectivity and content. Examines links among the nodes returned in a keyword-based query. Rankes the nodes, giving the highest rank to the most highly connected nodes. By doing so, finds hot spots on the Web that contain information germane to a user's query. WebQuery not only ranks and filters the results of a Web query; it also extends the result set beyond what the search engine retrieves, by finding interesting sites that are highly connected to those sites returned by the original query. Even with WebQuery filering and ranking query results, the result set can be enormous. Explores techniques for visualizing the returned information and discusses the criteria for using each of the technique
    Date
    1. 8.1996 22:08:06
    Footnote
    Contribution to a special issue of papers from the 6th International World Wide Web conference, held 7-11 Apr 1997, Santa Clara, California
    Source
    Computer networks and ISDN systems. 29(1997) no.8, S.1257-1267

Years

Languages

  • e 420
  • d 294
  • f 4
  • nl 4
  • ja 1
  • sp 1
  • More… Less…

Types

  • a 620
  • el 65
  • m 52
  • x 11
  • s 9
  • r 3
  • p 2
  • More… Less…

Subjects

Classifications