Search (1312 results, page 1 of 66)

  • × year_i:[2000 TO 2010}
  1. Lehrke, C.: Architektur von Suchmaschinen : Googles Architektur, insb. Crawler und Indizierer (2005) 0.11
    0.1131138 = product of:
      0.2827845 = sum of:
        0.24987002 = weight(_text_:crawler in 867) [ClassicSimilarity], result of:
          0.24987002 = score(doc=867,freq=4.0), product of:
            0.39420572 = queryWeight, product of:
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.048587184 = queryNorm
            0.6338569 = fieldWeight in 867, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.0390625 = fieldNorm(doc=867)
        0.032914463 = weight(_text_:22 in 867) [ClassicSimilarity], result of:
          0.032914463 = score(doc=867,freq=2.0), product of:
            0.17014404 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.048587184 = queryNorm
            0.19345059 = fieldWeight in 867, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0390625 = fieldNorm(doc=867)
      0.4 = coord(2/5)
    
    Abstract
    Das Internet mit seinen ständig neuen Usern und seinem extremen Wachstum bringt viele neue Herausforderungen mit sich. Aufgrund dieses Wachstums bedienen sich die meisten Leute der Hilfe von Suchmaschinen um Inhalte innerhalb des Internet zu finden. Suchmaschinen nutzen für die Beantwortung der User-Anfragen Information Retrieval Techniken. Problematisch ist nur, dass traditionelle Information Retrieval (IR) Systeme für eine relativ kleine und zusammenhängende Sammlung von Dokumenten entwickelt wurden. Das Internet hingegen unterliegt einem ständigen Wachstum, schnellen Änderungsraten und es ist über geographisch verteilte Computer verteilt. Aufgrund dieser Tatsachen müssen die alten Techniken erweitert oder sogar neue IRTechniken entwickelt werden. Eine Suchmaschine die diesen Herausforderungen vergleichsweise erfolgreich entgegnet ist Google. Ziel dieser Arbeit ist es aufzuzeigen, wie Suchmaschinen funktionieren. Der Fokus liegt dabei auf der Suchmaschine Google. Kapitel 2 wird sich zuerst mit dem Aufbau von Suchmaschinen im Allgemeinen beschäftigen, wodurch ein grundlegendes Verständnis für die einzelnen Komponenten geschaffen werden soll. Im zweiten Teil des Kapitels wird darauf aufbauend ein Überblick über die Architektur von Google gegeben. Kapitel 3 und 4 dienen dazu, näher auf die beiden Komponenten Crawler und Indexer einzugehen, bei denen es sich um zentrale Elemente im Rahmen von Suchmaschinen handelt.
    Pages
    22 S
  2. Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004) 0.11
    0.10840213 = product of:
      0.27100533 = sum of:
        0.23150799 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
          0.23150799 = score(doc=562,freq=2.0), product of:
            0.4119227 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.048587184 = queryNorm
            0.56201804 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
        0.039497353 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
          0.039497353 = score(doc=562,freq=2.0), product of:
            0.17014404 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.048587184 = queryNorm
            0.23214069 = fieldWeight in 562, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.046875 = fieldNorm(doc=562)
      0.4 = coord(2/5)
    
    Content
    Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
    Date
    8. 1.2013 10:22:32
  3. Reinke, S.; Schmidt, M.: Einmal suchen, alles finden : 7 Meta-Suchmaschinen im Test (2001) 0.07
    0.07067391 = product of:
      0.35336956 = sum of:
        0.35336956 = weight(_text_:crawler in 176) [ClassicSimilarity], result of:
          0.35336956 = score(doc=176,freq=2.0), product of:
            0.39420572 = queryWeight, product of:
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.048587184 = queryNorm
            0.89640903 = fieldWeight in 176, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.078125 = fieldNorm(doc=176)
      0.2 = coord(1/5)
    
    Abstract
    Von MetaSuchmaschinen oder Metacrawlern erwarten viele Datensucher Wunder. Die Crawler durchstöbern Kataloge von Suchmaschinen, fassen Ergebnisse zusammen, gleichen sie ab und präsentieren sie. CHIP hat sieben deutschsprachige, kostenlose Metacrawler getestet
  4. Schrodt, R.: Tiefen und Untiefen im wissenschaftlichen Sprachgebrauch (2008) 0.06
    0.061735462 = product of:
      0.30867732 = sum of:
        0.30867732 = weight(_text_:3a in 140) [ClassicSimilarity], result of:
          0.30867732 = score(doc=140,freq=2.0), product of:
            0.4119227 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.048587184 = queryNorm
            0.7493574 = fieldWeight in 140, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0625 = fieldNorm(doc=140)
      0.2 = coord(1/5)
    
    Content
    Vgl. auch: https://studylibde.com/doc/13053640/richard-schrodt. Vgl. auch: http%3A%2F%2Fwww.univie.ac.at%2FGermanistik%2Fschrodt%2Fvorlesung%2Fwissenschaftssprache.doc&usg=AOvVaw1lDLDR6NFf1W0-oC9mEUJf.
  5. Thelwall, M.: Results from a web impact factor crawler (2001) 0.06
    0.061205406 = product of:
      0.30602702 = sum of:
        0.30602702 = weight(_text_:crawler in 4490) [ClassicSimilarity], result of:
          0.30602702 = score(doc=4490,freq=6.0), product of:
            0.39420572 = queryWeight, product of:
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.048587184 = queryNorm
            0.776313 = fieldWeight in 4490, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4490)
      0.2 = coord(1/5)
    
    Abstract
    Web impact factors, the proposed web equivalent of impact factors for journals, can be calculated by using search engines. It has been found that the results are problematic because of the variable coverage of search engines as well as their ability to give significantly different results over short periods of time. The fundamental problem is that although some search engines provide a functionality that is capable of being used for impact calculations, this is not their primary task and therefore they do not give guarantees as to performance in this respect. In this paper, a bespoke web crawler designed specifically for the calculation of reliable WIFs is presented. This crawler was used to calculate WIFs for a number of UK universities, and the results of these calculations are discussed. The principal findings were that with certain restrictions, WIFs can be calculated reliably, but do not correlate with accepted research rankings owing to the variety of material hosted on university servers. Changes to the calculations to improve the fit of the results to research rankings are proposed, but there are still inherent problems undermining the reliability of the calculation. These problems still apply if the WIF scores are taken on their own as indicators of the general impact of any area of the Internet, but with care would not apply to online journals.
  6. Kwiatkowski, M.; Höhfeld, S.: Thematisches Aufspüren von Web-Dokumenten : eine kritische Betrachtung von Focused Crawling-Strategien (2007) 0.06
    0.061205406 = product of:
      0.30602702 = sum of:
        0.30602702 = weight(_text_:crawler in 153) [ClassicSimilarity], result of:
          0.30602702 = score(doc=153,freq=6.0), product of:
            0.39420572 = queryWeight, product of:
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.048587184 = queryNorm
            0.776313 = fieldWeight in 153, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.0390625 = fieldNorm(doc=153)
      0.2 = coord(1/5)
    
    Abstract
    Herkömmliche Suchmaschinen dienen der breiten Websuche und zeichnen sich zumeist durch eine hohe Quantität - nicht unbedingt durch Qualität - ihrer Ergebnismengen aus. Zum Suchen von Dokumenten wird ein allgemeiner Crawler eingesetzt, der Webseiten aufspürt, um große Datenspeicher aufzubauen. Focused Crawler gehen dagegen gezielter vor: Nicht enorme Datenmengen sollen durchsucht, gespeichert und indexiert werden, sondern nur bestimmte, thematisch relevante Segmente des World Wide Web. Der Focused Crawler muss einen möglichst optimalen Weg durch das Web finden, um Knowledge Discovery zu betreiben. Dabei bleiben die für eine Thematik irrelevanten Bereiche des Web unberücksichtigt. Die Aufgabe wird dadurch erheblich verkleinert und der Ressourcenaufwand verringert. Ziel ist die Produktion qualifizierter Suchergebnisse zu einem bestimmten Wissensgebiet. Im Allgemeinen können Focused Crawling-Techniken für den Aufbau spezialisierter vertikaler Suchmaschinen eingesetzt werden. Sie sind darüber hinaus im Bereich der Digitalen Bibliotheken von Vorteil. Da diese oft über einen thematischen Schwerpunkt verfügen und der qualifizierten Literatur-Untersuchung dienen, müssen sie einen gewissen Qualitätsanspruch Genüge leisten und dabei lediglich Anfragen zu einem definierten Wissensbereich bedienen. Der Einsatz von Focused Crawling bietet sich also an, um eine hohe Dokument-Qualität in einer spezifischen Domäne zu gewährleisten. Dieser Review-Artikel beleuchtet grundlegende Ansätze des Focused Crawling und verfolgt diese bis in die aktuellen Entwicklungen. Praktische Einsatzgebiete und aktuelle Systeme untermauern die Bedeutsamkeit des Forschungsgebiets. Darüber hinaus wird eine kritische Betrachtung der aufgeführten Ansätze geleistet.
  7. Jung, J.J.: Contextualized query sampling to discover semantic resource descriptions on the web (2009) 0.06
    0.0599688 = product of:
      0.299844 = sum of:
        0.299844 = weight(_text_:crawler in 4216) [ClassicSimilarity], result of:
          0.299844 = score(doc=4216,freq=4.0), product of:
            0.39420572 = queryWeight, product of:
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.048587184 = queryNorm
            0.7606282 = fieldWeight in 4216, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.046875 = fieldNorm(doc=4216)
      0.2 = coord(1/5)
    
    Abstract
    Resource description extracted by query-sampling method can be applied to determine which database sources a certain query should be firstly sent to. In this paper, we propose a contextualized query-sampling method to extract the resources which are most relevant to up-to-date context. Practically, the proposed approach is adopted to personal crawler systems (the so-called focused crawlers), which can support the corresponding user's web navigation tasks in real-time. By taking into account the user context (e.g., intentions or interests), the crawler can build the queries to evaluate candidate information sources. As a result, we can discover semantic associations (i) between user context and the sources, and (ii) between all pairs of the sources. These associations are applied to rank the sources, and transform the queries for the other sources. For evaluating the performance of contextualized query sampling on 53 information sources, we compared the ranking lists recommended by the proposed method with user feedbacks (i.e., ideal ranks), and also computed the precision of discovered subsumptions as semantic associations between the sources.
  8. Reibold, H.: Findigkeit gefragt (2000) 0.06
    0.05653913 = product of:
      0.28269565 = sum of:
        0.28269565 = weight(_text_:crawler in 4283) [ClassicSimilarity], result of:
          0.28269565 = score(doc=4283,freq=2.0), product of:
            0.39420572 = queryWeight, product of:
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.048587184 = queryNorm
            0.7171272 = fieldWeight in 4283, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.0625 = fieldNorm(doc=4283)
      0.2 = coord(1/5)
    
    Object
    Crawler
  9. Vetere, G.; Lenzerini, M.: Models for semantic interoperability in service-oriented architectures (2005) 0.05
    0.054018535 = product of:
      0.27009267 = sum of:
        0.27009267 = weight(_text_:3a in 306) [ClassicSimilarity], result of:
          0.27009267 = score(doc=306,freq=2.0), product of:
            0.4119227 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.048587184 = queryNorm
            0.65568775 = fieldWeight in 306, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0546875 = fieldNorm(doc=306)
      0.2 = coord(1/5)
    
    Content
    Vgl.: http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=5386707&url=http%3A%2F%2Fieeexplore.ieee.org%2Fxpls%2Fabs_all.jsp%3Farnumber%3D5386707.
  10. Stock, M.; Stock, W.G.: Internet-Suchwerkzeuge im Vergleich : Teil 1: Retrievaltests mit Known Item searches (2000) 0.05
    0.049471736 = product of:
      0.24735868 = sum of:
        0.24735868 = weight(_text_:crawler in 5772) [ClassicSimilarity], result of:
          0.24735868 = score(doc=5772,freq=2.0), product of:
            0.39420572 = queryWeight, product of:
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.048587184 = queryNorm
            0.6274863 = fieldWeight in 5772, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5772)
      0.2 = coord(1/5)
    
    Object
    Web-Crawler
  11. Thelwall, M.; Stuart, D.: Web crawling ethics revisited : cost, privacy, and denial of service (2006) 0.05
    0.049471736 = product of:
      0.24735868 = sum of:
        0.24735868 = weight(_text_:crawler in 6098) [ClassicSimilarity], result of:
          0.24735868 = score(doc=6098,freq=2.0), product of:
            0.39420572 = queryWeight, product of:
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.048587184 = queryNorm
            0.6274863 = fieldWeight in 6098, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.0546875 = fieldNorm(doc=6098)
      0.2 = coord(1/5)
    
    Abstract
    Ethical aspects of the employment of Web crawlers for information science research and other contexts are reviewed. The difference between legal and ethical uses of communications technologies is emphasized as well as the changing boundary between ethical and unethical conduct. A review of the potential impacts on Web site owners is used to underpin a new framework for ethical crawling, and it is argued that delicate human judgment is required for each individual case, with verdicts likely to change over time. Decisions can be based upon an approximate cost-benefit analysis, but it is crucial that crawler owners find out about the technological issues affecting the owners of the sites being crawled in order to produce an informed assessment.
  12. Mas, S.; Marleau, Y.: Proposition of a faceted classification model to support corporate information organization and digital records management (2009) 0.05
    0.0463016 = product of:
      0.23150799 = sum of:
        0.23150799 = weight(_text_:3a in 2918) [ClassicSimilarity], result of:
          0.23150799 = score(doc=2918,freq=2.0), product of:
            0.4119227 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.048587184 = queryNorm
            0.56201804 = fieldWeight in 2918, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=2918)
      0.2 = coord(1/5)
    
    Footnote
    Vgl.: http://ieeexplore.ieee.org/Xplore/login.jsp?reload=true&url=http%3A%2F%2Fieeexplore.ieee.org%2Fiel5%2F4755313%2F4755314%2F04755480.pdf%3Farnumber%3D4755480&authDecision=-203.
  13. Wenyin, L.; Chen, Z.; Li, M.; Zhang, H.: ¬A media agent for automatically builiding a personalized semantic index of Web media objects (2001) 0.04
    0.042404346 = product of:
      0.21202172 = sum of:
        0.21202172 = weight(_text_:crawler in 6522) [ClassicSimilarity], result of:
          0.21202172 = score(doc=6522,freq=2.0), product of:
            0.39420572 = queryWeight, product of:
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.048587184 = queryNorm
            0.5378454 = fieldWeight in 6522, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.046875 = fieldNorm(doc=6522)
      0.2 = coord(1/5)
    
    Abstract
    A novel idea of media agent is briefly presented, which can automatically build a personalized semantic index of Web media objects for each particular user. Because the Web is a rich source of multimedia data and the text content on the Web pages is usually semantically related to those media objects on the same pages, the media agent can automatically collect the URLs and related text, and then build the index of the multimedia data, on behalf of the user whenever and wherever she accesses these multimedia data or their container Web pages. Moreover, the media agent can also use an off-line crawler to build the index for those multimedia objects that are relevant to the user's favorites but have not accessed by the user yet. When the user wants to find these multimedia data once again, the semantic index facilitates text-based search for her.
  14. Kaiser, C.: Mit "Neomo" und "Turbo 10" neue Initiativen auf dem deutschen und britischen Suchmarkt (2005) 0.04
    0.042404346 = product of:
      0.21202172 = sum of:
        0.21202172 = weight(_text_:crawler in 3434) [ClassicSimilarity], result of:
          0.21202172 = score(doc=3434,freq=2.0), product of:
            0.39420572 = queryWeight, product of:
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.048587184 = queryNorm
            0.5378454 = fieldWeight in 3434, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.046875 = fieldNorm(doc=3434)
      0.2 = coord(1/5)
    
    Abstract
    "Search Engine Strategies Conference" (SES) in München mit 160 Teilnehmern. Die Vortragenden waren im Regelfall sehr kompetent, und die Zuhörerschaft schien gut vorinformiert zu sein. Trotzdem wäre bei manchen Vorträgen mehr Inhalt und Fachkompetenz wünschenswert gewesen - wie beispielsweise beim Vortrag von Google. Die geplante Session "Treffen Sie die Crawler" fand leider nicht statt. Mittlerweile gibt es andere interessante Konferenzen in Europas, die sich mit Suchmaschinenmarketing und -optimierung befassten, wie das "Suchmaschinenmarketingseminar" in Heidelberg im November 2004, das wenig besucht war, aber hochinteressante Fachvorträge und Diskussionsforen bot. Die SES gilt bisher als das wichtigste Branchenereignis für Suchmaschinenmarketing und -optimierung weltweit. Hier treffen sich Websiteanbieter, Suchmaschinenmarketingagenturen und Suchmaschinenbetreiber. Außer allgemeinen Einblicken in die aktuelle Entwicklung der Branche bietet die SES Informationen zu Themen wie dynamische Websites, Websitestruktur, Verlinkung und Keywordanalysen. Neue Themen waren "lokale Suche", die aktuelle Entwicklung im deutschen Suchmarkt und markenrechtliche Probleme. Websiteanbieter konnten in den "Website-Klinik"-Sessions ihre Sites von Experten prüfen lassen und praktische Tipps zur Verbesserung des Rankings einholen.
  15. Thelwall, M.: Text characteristics of English language university Web sites (2005) 0.04
    0.042404346 = product of:
      0.21202172 = sum of:
        0.21202172 = weight(_text_:crawler in 3463) [ClassicSimilarity], result of:
          0.21202172 = score(doc=3463,freq=2.0), product of:
            0.39420572 = queryWeight, product of:
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.048587184 = queryNorm
            0.5378454 = fieldWeight in 3463, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.046875 = fieldNorm(doc=3463)
      0.2 = coord(1/5)
    
    Abstract
    The nature of the contents of academic Web sites is of direct relevance to the new field of scientific Web intelligence, and for search engine and topic-specific crawler designers. We analyze word frequencies in national academic Webs using the Web sites of three Englishspeaking nations: Australia, New Zealand, and the United Kingdom. Strong regularities were found in page size and word frequency distributions, but with significant anomalies. At least 26% of pages contain no words. High frequency words include university names and acronyms, Internet terminology, and computing product names: not always words in common usage away from the Web. A minority of low frequency words are spelling mistakes, with other common types including nonwords, proper names, foreign language terms or computer science variable names. Based upon these findings, recommendations for data cleansing and filtering are made, particularly for clustering applications.
  16. Ding, Y.; Jacob, E.K.; Zhang, Z.; Foo, S.; Yan, E.; George, N.L.; Guo, L.: Perspectives on social tagging (2009) 0.04
    0.042404346 = product of:
      0.21202172 = sum of:
        0.21202172 = weight(_text_:crawler in 3290) [ClassicSimilarity], result of:
          0.21202172 = score(doc=3290,freq=2.0), product of:
            0.39420572 = queryWeight, product of:
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.048587184 = queryNorm
            0.5378454 = fieldWeight in 3290, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.046875 = fieldNorm(doc=3290)
      0.2 = coord(1/5)
    
    Abstract
    Social tagging is one of the major phenomena transforming the World Wide Web from a static platform into an actively shared information space. This paper addresses various aspects of social tagging, including different views on the nature of social tagging, how to make use of social tags, and how to bridge social tagging with other Web functionalities; it discusses the use of facets to facilitate browsing and searching of tagging data; and it presents an analogy between bibliometrics and tagometrics, arguing that established bibliometric methodologies can be applied to analyze tagging behavior on the Web. Based on the Upper Tag Ontology (UTO), a Web crawler was built to harvest tag data from Delicious, Flickr, and YouTube in September 2007. In total, 1.8 million objects, including bookmarks, photos, and videos, 3.1 million taggers, and 12.1 million tags were collected and analyzed. Some tagging patterns and variations are identified and discussed.
  17. Ding, L.; Finin, T.; Joshi, A.; Peng, Y.; Cost, R.S.; Sachs, J.; Pan, R.; Reddivari, P.; Doshi, V.: Swoogle : a Semantic Web search and metadata engine (2004) 0.04
    0.042404346 = product of:
      0.21202172 = sum of:
        0.21202172 = weight(_text_:crawler in 4704) [ClassicSimilarity], result of:
          0.21202172 = score(doc=4704,freq=2.0), product of:
            0.39420572 = queryWeight, product of:
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.048587184 = queryNorm
            0.5378454 = fieldWeight in 4704, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.046875 = fieldNorm(doc=4704)
      0.2 = coord(1/5)
    
    Abstract
    Swoogle is a crawler-based indexing and retrieval system for the Semantic Web, i.e., for Web documents in RDF or OWL. It extracts metadata for each discovered document, and computes relations between documents. Discovered documents are also indexed by an information retrieval system which can use either character N-Gram or URIrefs as keywords to find relevant documents and to compute the similarity among a set of documents. One of the interesting properties we compute is rank, a measure of the importance of a Semantic Web document.
  18. Donsbach, W.: Wahrheit in den Medien : über den Sinn eines methodischen Objektivitätsbegriffes (2001) 0.04
    0.038584664 = product of:
      0.19292332 = sum of:
        0.19292332 = weight(_text_:3a in 5895) [ClassicSimilarity], result of:
          0.19292332 = score(doc=5895,freq=2.0), product of:
            0.4119227 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.048587184 = queryNorm
            0.46834838 = fieldWeight in 5895, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5895)
      0.2 = coord(1/5)
    
    Source
    Politische Meinung. 381(2001) Nr.1, S.65-74 [https%3A%2F%2Fwww.dgfe.de%2Ffileadmin%2FOrdnerRedakteure%2FSektionen%2FSek02_AEW%2FKWF%2FPublikationen_Reihe_1989-2003%2FBand_17%2FBd_17_1994_355-406_A.pdf&usg=AOvVaw2KcbRsHy5UQ9QRIUyuOLNi]
  19. Ackermann, E.: Piaget's constructivism, Papert's constructionism : what's the difference? (2001) 0.04
    0.038584664 = product of:
      0.19292332 = sum of:
        0.19292332 = weight(_text_:3a in 692) [ClassicSimilarity], result of:
          0.19292332 = score(doc=692,freq=2.0), product of:
            0.4119227 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.048587184 = queryNorm
            0.46834838 = fieldWeight in 692, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=692)
      0.2 = coord(1/5)
    
    Content
    Vgl.: https://www.semanticscholar.org/paper/Piaget-%E2%80%99-s-Constructivism-%2C-Papert-%E2%80%99-s-%3A-What-%E2%80%99-s-Ackermann/89cbcc1e740a4591443ff4765a6ae8df0fdf5554. Darunter weitere Hinweise auf verwandte Beiträge. Auch unter: Learning Group Publication 5(2001) no.3, S.438.
  20. Becker, A: Neue Suchmaschinen für Fortgeschrittene : Neue Such-Angebote: Die fünf Top-Newcomer im Überblick (2000) 0.04
    0.035336956 = product of:
      0.17668478 = sum of:
        0.17668478 = weight(_text_:crawler in 1526) [ClassicSimilarity], result of:
          0.17668478 = score(doc=1526,freq=2.0), product of:
            0.39420572 = queryWeight, product of:
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.048587184 = queryNorm
            0.44820452 = fieldWeight in 1526, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.113368 = idf(docFreq=35, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1526)
      0.2 = coord(1/5)
    
    Content
    Kart00.com - Anstatt seine Resultate traditionell als Liste darzustellen, zeichnet der Meta-Sucher eine Ergebniskarte. Vorteil: Die bildliche Darstellung sorgtfür einen überzeugenden Themenüberblick. Teoma.com - Die Maschine fahndet mit drei unterschiedlichen Methoden: via Volltextsuche, über Expertenseiten und mithilfe von Schlagwörtern. Vorteil: Durch die innovative 3D-Suche erzielt Teoma bei speziellen Recherchen beachtliche Ergebnisse. Wondir.com - Zu jeder Anfrage gibt es bei Wondir Antworten auf fünf Ebenen. Von einer Trefferliste bis hin zu einem Experten-MailKontakt. Vorteil: ideal für komplizierte und wissenschaftliche Themen. Turb10.com - Der neue britische Meta-Crawler durchforstet gleichzeitig sowohl das normale als auch das Deep Web. Vorteil: Dank Turb10.com muss niemand mehr für Deep-Web-Recherchen auf spezielle Zusatzprogramme zurückgreifen. Hotbot.com - Der Ex-Volitextdienst setzt jetzt auf Service. Über seine Startseite lassen sich vier Top-Dienste (u.a. Google, FAST) abfragen. Vorteil: Hotbot bietet vier Spitzenangebote auf einen Klick.

Languages

Types

  • a 1106
  • m 141
  • el 63
  • s 49
  • b 26
  • x 13
  • i 8
  • n 2
  • r 1
  • More… Less…

Themes

Subjects

Classifications