Search (214 results, page 1 of 11)

  • × year_i:[2020 TO 2030}
  1. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.21
    0.20833743 = product of:
      0.41667485 = sum of:
        0.05952498 = product of:
          0.17857493 = sum of:
            0.17857493 = weight(_text_:3a in 862) [ClassicSimilarity], result of:
              0.17857493 = score(doc=862,freq=2.0), product of:
                0.3177388 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03747799 = queryNorm
                0.56201804 = fieldWeight in 862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.046875 = fieldNorm(doc=862)
          0.33333334 = coord(1/3)
        0.17857493 = weight(_text_:2f in 862) [ClassicSimilarity], result of:
          0.17857493 = score(doc=862,freq=2.0), product of:
            0.3177388 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03747799 = queryNorm
            0.56201804 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
        0.17857493 = weight(_text_:2f in 862) [ClassicSimilarity], result of:
          0.17857493 = score(doc=862,freq=2.0), product of:
            0.3177388 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03747799 = queryNorm
            0.56201804 = fieldWeight in 862, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.046875 = fieldNorm(doc=862)
      0.5 = coord(3/6)
    
    Source
    https%3A%2F%2Farxiv.org%2Fabs%2F2212.06721&usg=AOvVaw3i_9pZm9y_dQWoHi6uv0EN
  2. Gabler, S.: Vergabe von DDC-Sachgruppen mittels eines Schlagwort-Thesaurus (2021) 0.17
    0.17361453 = product of:
      0.34722906 = sum of:
        0.049604155 = product of:
          0.14881246 = sum of:
            0.14881246 = weight(_text_:3a in 1000) [ClassicSimilarity], result of:
              0.14881246 = score(doc=1000,freq=2.0), product of:
                0.3177388 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03747799 = queryNorm
                0.46834838 = fieldWeight in 1000, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1000)
          0.33333334 = coord(1/3)
        0.14881246 = weight(_text_:2f in 1000) [ClassicSimilarity], result of:
          0.14881246 = score(doc=1000,freq=2.0), product of:
            0.3177388 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03747799 = queryNorm
            0.46834838 = fieldWeight in 1000, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1000)
        0.14881246 = weight(_text_:2f in 1000) [ClassicSimilarity], result of:
          0.14881246 = score(doc=1000,freq=2.0), product of:
            0.3177388 = queryWeight, product of:
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.03747799 = queryNorm
            0.46834838 = fieldWeight in 1000, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.478011 = idf(docFreq=24, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1000)
      0.5 = coord(3/6)
    
    Content
    Master thesis Master of Science (Library and Information Studies) (MSc), Universität Wien. Advisor: Christoph Steiner. Vgl.: https://www.researchgate.net/publication/371680244_Vergabe_von_DDC-Sachgruppen_mittels_eines_Schlagwort-Thesaurus. DOI: 10.25365/thesis.70030. Vgl. dazu die Präsentation unter: https://www.google.com/url?sa=i&rct=j&q=&esrc=s&source=web&cd=&ved=0CAIQw7AJahcKEwjwoZzzytz_AhUAAAAAHQAAAAAQAg&url=https%3A%2F%2Fwiki.dnb.de%2Fdownload%2Fattachments%2F252121510%2FDA3%2520Workshop-Gabler.pdf%3Fversion%3D1%26modificationDate%3D1671093170000%26api%3Dv2&psig=AOvVaw0szwENK1or3HevgvIDOfjx&ust=1687719410889597&opi=89978449.
  3. Meineck, S.: Gesichter-Suchmaschine PimEyes bricht das Schweigen : Neuer Chef (2022) 0.04
    0.03821566 = product of:
      0.22929396 = sum of:
        0.22929396 = weight(_text_:suchmaschine in 418) [ClassicSimilarity], result of:
          0.22929396 = score(doc=418,freq=6.0), product of:
            0.21191008 = queryWeight, product of:
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03747799 = queryNorm
            1.0820342 = fieldWeight in 418, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.078125 = fieldNorm(doc=418)
      0.16666667 = coord(1/6)
    
    Abstract
    PimEyes untergräbt die Anonymität von Menschen, deren Gesicht im Internet zu finden ist. Nach breiter Kritik hatte sich die polnische Suchmaschine auf die Seychellen abgesetzt. Jetzt hat PimEyes einen neuen Chef - und geht an die Öfffentlichkeit.
    Source
    https://netzpolitik.org/2022/neuer-chef-gesichter-suchmaschine-pimeyes-bricht-das-schweigen/?utm_source=pocket-newtab-global-de-DE
  4. Amirhosseini, M.: ¬A novel method for ranking knowledge organization systems (KOSs) based on cognition states (2022) 0.03
    0.034583665 = product of:
      0.10375099 = sum of:
        0.0969192 = weight(_text_:ranking in 1105) [ClassicSimilarity], result of:
          0.0969192 = score(doc=1105,freq=8.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.47809508 = fieldWeight in 1105, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03125 = fieldNorm(doc=1105)
        0.0068317903 = product of:
          0.02049537 = sum of:
            0.02049537 = weight(_text_:29 in 1105) [ClassicSimilarity], result of:
              0.02049537 = score(doc=1105,freq=2.0), product of:
                0.13183585 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03747799 = queryNorm
                0.15546128 = fieldWeight in 1105, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1105)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    The purpose of this article is to delineate the process of evolution of know­ledge organization systems (KOSs) through identification of principles of unity such as internal and external unity in organizing the structure of KOSs to achieve content storage and retrieval purposes and to explain a novel method used in ranking of KOSs by proposing the principle of rank unity. Different types of KOSs which are addressed in this article include dictionaries, Roget's thesaurus, thesauri, micro, macro, and meta-thesaurus, ontologies, and lower, middle, and upper-level ontologies. This article relied on dialectic models to clarify the ideas in Kant's know­ledge theory. This is done by identifying logical relationships between categories (i.e., Thesis, antithesis, and synthesis) in the creation of data, information, and know­ledge in the human mind. The Analysis has adapted a historical methodology, more specifically a documentary method, as its reasoning process to propose a conceptual model for ranking KOSs. The study endeavors to explain the main elements of data, information, and know­ledge along with engineering mechanisms such as data, information, and know­ledge engineering in developing the structure of KOSs and also aims to clarify their influence on content storage and retrieval performance. KOSs have followed related principles of order to achieve an internal order, which could be examined by analyzing the principle of internal unity in know­ledge organizations. The principle of external unity leads us to the necessity of compatibility and interoperability between different types of KOSs to achieve semantic harmonization in increasing the performance of content storage and retrieval. Upon introduction of the principle of rank unity, a ranking method of KOSs utilizing cognition states as criteria could be considered to determine the position of each know­ledge organization with respect to others. The related criteria of the principle of rank unity- cognition states- are derived from Immanuel Kant's epistemology. The research results showed that KOSs, while having defined positions in cognition states, specific principles of order, related operational mechanisms, and related principles of unity in achieving their specific purposes, have benefited from the developmental experiences of previous KOSs, and further, their developmental processes owe to the experiences and methods of their previous generations.
    Date
    19.11.2023 19:07:29
  5. Weiß, E.-M.: ChatGPT soll es richten : Microsoft baut KI in Suchmaschine Bing ein (2023) 0.03
    0.03453535 = product of:
      0.20721208 = sum of:
        0.20721208 = weight(_text_:suchmaschine in 866) [ClassicSimilarity], result of:
          0.20721208 = score(doc=866,freq=10.0), product of:
            0.21191008 = queryWeight, product of:
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03747799 = queryNorm
            0.9778302 = fieldWeight in 866, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.0546875 = fieldNorm(doc=866)
      0.16666667 = coord(1/6)
    
    Abstract
    ChatGPT, die künstliche Intelligenz der Stunde, ist von OpenAI entwickelt worden. Und OpenAI ist in der Vergangenheit nicht unerheblich von Microsoft unterstützt worden. Nun geht es ums Profitieren: Die KI soll in die Suchmaschine Bing eingebaut werden, was eine direkte Konkurrenz zu Googles Suchalgorithmen und Intelligenzen bedeutet. Bing war da bislang nicht sonderlich erfolgreich. Wie "The Information" mit Verweis auf zwei Insider berichtet, plant Microsoft, ChatGPT in seine Suchmaschine Bing einzubauen. Bereits im März könnte die neue, intelligente Suche verfügbar sein. Microsoft hatte zuvor auf der hauseigenen Messe Ignite zunächst die Integration des Bildgenerators DALL·E 2 in seine Suchmaschine angekündigt - ohne konkretes Startdatum jedoch. Fragt man ChatGPT selbst, bestätigt der Chatbot seine künftige Aufgabe noch nicht. Weiß aber um potentielle Vorteile.
    Source
    https://www.heise.de/news/ChatGPT-soll-es-richten-Microsoft-baut-KI-in-Suchmaschine-Bing-ein-7447837.html
  6. Dang, E.K.F.; Luk, R.W.P.; Allan, J.: ¬A retrieval model family based on the probability ranking principle for ad hoc retrieval (2022) 0.02
    0.024480894 = product of:
      0.14688537 = sum of:
        0.14688537 = weight(_text_:ranking in 638) [ClassicSimilarity], result of:
          0.14688537 = score(doc=638,freq=6.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.7245744 = fieldWeight in 638, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.0546875 = fieldNorm(doc=638)
      0.16666667 = coord(1/6)
    
    Abstract
    Many successful retrieval models are derived based on or conform to the probability ranking principle (PRP). We present a new derivation of a document ranking function given by the probability of relevance of a document, conforming to the PRP. Our formulation yields a family of retrieval models, called probabilistic binary relevance (PBR) models, with various instantiations obtained by different probability estimations. By extensive experiments on a range of TREC collections, improvement of the PBR models over some established baselines with statistical significance is observed, especially in the large Clueweb09 Cat-B collection.
  7. Option für Metager als Standardsuchmaschine, Suchmaschine nach dem Peer-to-Peer-Prinzip (2021) 0.02
    0.023350157 = product of:
      0.14010094 = sum of:
        0.14010094 = weight(_text_:suchmaschine in 431) [ClassicSimilarity], result of:
          0.14010094 = score(doc=431,freq=14.0), product of:
            0.21191008 = queryWeight, product of:
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03747799 = queryNorm
            0.6611339 = fieldWeight in 431, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03125 = fieldNorm(doc=431)
      0.16666667 = coord(1/6)
    
    Content
    "Option für Metager als Standardsuchmaschine. Google wurde von der EU verordnet, auf Android-Smartphones bei Neukonfiguration eine Auswahl an Suchmaschinen anzubieten, die als Standardsuchmaschine eingerichtet werden können. Suchmaschinen konnten sich im Rahmen einer Auktion bewerben. Auch wir hatten am Auktionsverfahren teilgenommen, jedoch rein formell mit einem Gebot von null Euro. Nun wurde Google von der EU angewiesen, auf das wettbewerbsverzerrende Auktionsverfahren zu verzichten und alle angemeldeten Suchmaschinen als Option anzubieten. Auf Android ist es nun optional möglich, MetaGer als Standardsuchmaschine für den Bereich D/A/CH auszuwählen. Zwar werden nicht immer alle Suchmaschinen zur Auswahl angezeigt, aber das Zufallsprinzip sorgt immerhin dafür, dass jede Suchmaschine mit einer gewissen Wahrscheinlichkeit in der Liste zu finden ist.
    Auch auf dem Volla-Phone ist es bald möglich, MetaGer als Standardsuchmaschine zu wählen. Das Volla Phone ist ein Produkt von "Hallo Welt Systeme UG" in Remscheid. Die Entwickler des Smartphones verfolgen den Ansatz, möglichst wenig von der Aufmerksamkeit des Nutzers zu beanspruchen. Technik soll nicht ablenken und sich in der Vordergrund spielen, sondern als bloßes Werkzeug im Hintergrund bleiben. Durch Möglichkeiten wie detaillierter Datenschutzeinstellungen, logfreiem VPN, quelloffener Apps aus einem alternativen App Store wird zudem Schutz der Privatsphäre ermöglicht - ganz ohne Google-Dienste. Durch die Partnerschaft mit MetaGer können die Nutzer von Volla-Phone auch im Bereich Suchmaschine Privatsphärenschutz realisieren. Mehr unter: https://suma-ev.de/mit-metager-auf-dem-volla-phone-suchen/
    YaCy: Suchmaschine nach dem Peer-to-Peer-Prinzip. YaCy ist eine dezentrale, freie Suchmaschine. Die Besonderheit: die freie Suchmaschine läuft nicht auf zentralen Servern eines einzelnen Betreibers, sondern funktioniert nach dem Peer-to-Peer (P2P) Prinzip. Dieses basiert darauf, dass die YaCy-Nutzer aufgerufene Webseiten auf ihrem Computer lokal indexieren. Jeder Nutzer "ercrawlt" sich damit einen kleinen Index, den er durch Kommunikation mit anderen YaCy-Peers teilen kann. Das Programm sorgt dafür, dass durch die kleinen dezentralen Crawler einzelner Nutzer schließlich ein globaler Gesamtindex entsteht. Je mehr Nutzer Teil dieser dezentralen Suche sind, desto größer wird der gemeinsame Index, auf den der einzelne Nutzer dann Zugriff haben kann. Seit kurzem befindet sich YaCy im Verbund unserer abgefragten Suchmaschinen. Wir sind somit auch Teil des Indexes der Suchmaschine.
  8. Wang, J.; Halffman, W.; Zhang, Y.H.: Sorting out journals : the proliferation of journal lists in China (2023) 0.02
    0.023012474 = product of:
      0.06903742 = sum of:
        0.0605745 = weight(_text_:ranking in 1055) [ClassicSimilarity], result of:
          0.0605745 = score(doc=1055,freq=2.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.29880944 = fieldWeight in 1055, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1055)
        0.008462917 = product of:
          0.025388751 = sum of:
            0.025388751 = weight(_text_:22 in 1055) [ClassicSimilarity], result of:
              0.025388751 = score(doc=1055,freq=2.0), product of:
                0.13124153 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03747799 = queryNorm
                0.19345059 = fieldWeight in 1055, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1055)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    Journal lists are instruments to categorize, compare, and assess research and scholarly publications. Our study investigates the remarkable proliferation of such journal lists in China, analyses their underlying values, quality criteria and ranking principles, and specifies how concerns specific to the Chinese research policy and publishing system inform these lists. Discouraged lists of "bad journals" reflect concerns over inferior research publications, but also the involved drain on public resources. Endorsed lists of "good journals" are based on criteria valued in research policy, reflecting the distinctive administrative logic of state-led Chinese research and publishing policy, ascribing worth to scientific journals for its specific national and institutional needs. In this regard, the criteria used for journal list construction are contextual and reflect the challenges of public resource allocation in a market-led publication system. Chinese journal lists therefore reflect research policy changes, such as a shift away from output-dominated research evaluation, the specific concerns about research misconduct, and balancing national research needs against international standards, resulting in distinctly Chinese quality criteria. However, contrasting concerns and inaccuracies lead to contradictions in the "qualify" and "disqualify" binary logic and demonstrate inherent tensions and limitations in journal lists as policy tools.
    Date
    22. 9.2023 16:39:23
  9. Luo, L.; Ju, J.; Li, Y.-F.; Haffari, G.; Xiong, B.; Pan, S.: ChatRule: mining logical rules with large language models for knowledge graph reasoning (2023) 0.02
    0.023012474 = product of:
      0.06903742 = sum of:
        0.0605745 = weight(_text_:ranking in 1171) [ClassicSimilarity], result of:
          0.0605745 = score(doc=1171,freq=2.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.29880944 = fieldWeight in 1171, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1171)
        0.008462917 = product of:
          0.025388751 = sum of:
            0.025388751 = weight(_text_:22 in 1171) [ClassicSimilarity], result of:
              0.025388751 = score(doc=1171,freq=2.0), product of:
                0.13124153 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03747799 = queryNorm
                0.19345059 = fieldWeight in 1171, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1171)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Abstract
    Logical rules are essential for uncovering the logical connections between relations, which could improve the reasoning performance and provide interpretable results on knowledge graphs (KGs). Although there have been many efforts to mine meaningful logical rules over KGs, existing methods suffer from the computationally intensive searches over the rule space and a lack of scalability for large-scale KGs. Besides, they often ignore the semantics of relations which is crucial for uncovering logical connections. Recently, large language models (LLMs) have shown impressive performance in the field of natural language processing and various applications, owing to their emergent ability and generalizability. In this paper, we propose a novel framework, ChatRule, unleashing the power of large language models for mining logical rules over knowledge graphs. Specifically, the framework is initiated with an LLM-based rule generator, leveraging both the semantic and structural information of KGs to prompt LLMs to generate logical rules. To refine the generated rules, a rule ranking module estimates the rule quality by incorporating facts from existing KGs. Last, a rule validator harnesses the reasoning ability of LLMs to validate the logical correctness of ranked rules through chain-of-thought reasoning. ChatRule is evaluated on four large-scale KGs, w.r.t. different rule quality metrics and downstream tasks, showing the effectiveness and scalability of our method.
    Date
    23.11.2023 19:07:22
  10. Lewandowski, D.: Suchmaschinen (2023) 0.02
    0.022929398 = product of:
      0.13757639 = sum of:
        0.13757639 = weight(_text_:suchmaschine in 793) [ClassicSimilarity], result of:
          0.13757639 = score(doc=793,freq=6.0), product of:
            0.21191008 = queryWeight, product of:
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03747799 = queryNorm
            0.6492206 = fieldWeight in 793, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.046875 = fieldNorm(doc=793)
      0.16666667 = coord(1/6)
    
    Abstract
    Eine Suchmaschine (auch: Web-Suchmaschine, Universalsuchmaschine) ist ein Computersystem, das Inhalte aus dem World Wide Web (WWW) mittels Crawling erfasst und über eine Benutzerschnittstelle durchsuchbar macht, wobei die Ergebnisse in einer nach systemseitig angenommener Relevanz geordneten Darstellung aufgeführt werden. Dies bedeutet, dass Suchmaschinen im Gegensatz zu anderen Informationssystemen nicht auf einem klar abgegrenzten Datenbestand aufbauen, sondern diesen aus den verstreut vorliegenden Dokumenten des WWW zusammenstellen. Dieser Datenbestand wird über eine Benutzerschnittstelle zugänglich gemacht, die so gestaltet ist, dass die Suchmaschine von Laien problemlos genutzt werden kann. Die zu einer Suchanfrage ausgegebenen Treffer werden so sortiert, dass den Nutzenden die aus Systemsicht relevantesten Dokumente zuerst angezeigt werden. Dabei handelt es sich um komplexe Bewertungsverfahren, denen zahlreiche Annahmen über die Relevanz von Dokumenten in Bezug auf Suchanfragen zugrunde liegen.
  11. Wiggers, G.; Verberne, S.; Loon, W. van; Zwenne, G.-J.: Bibliometric-enhanced legal information retrieval : combining usage and citations as flavors of impact relevance (2023) 0.02
    0.022574786 = product of:
      0.13544871 = sum of:
        0.13544871 = weight(_text_:ranking in 1022) [ClassicSimilarity], result of:
          0.13544871 = score(doc=1022,freq=10.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.66815823 = fieldWeight in 1022, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1022)
      0.16666667 = coord(1/6)
    
    Abstract
    Bibliometric-enhanced information retrieval uses bibliometrics (e.g., citations) to improve ranking algorithms. Using a data-driven approach, this article describes the development of a bibliometric-enhanced ranking algorithm for legal information retrieval, and the evaluation thereof. We statistically analyze the correlation between usage of documents and citations over time, using data from a commercial legal search engine. We then propose a bibliometric boost function that combines usage of documents with citation counts. The core of this function is an impact variable based on usage and citations that increases in influence as citations and usage counts become more reliable over time. We evaluate our ranking function by comparing search sessions before and after the introduction of the new ranking in the search engine. Using a cost model applied to 129,571 sessions before and 143,864 sessions after the intervention, we show that our bibliometric-enhanced ranking algorithm reduces the time of a search session of legal professionals by 2 to 3% on average for use cases other than known-item retrieval or updating behavior. Given the high hourly tariff of legal professionals and the limited time they can spend on research, this is expected to lead to increased efficiency, especially for users with extremely long search sessions.
  12. Lobo, S.: ¬Das Ende von Google, wie wir es kannten : Bessere Treffer durch ChatGPT (2022) 0.02
    0.022063822 = product of:
      0.13238293 = sum of:
        0.13238293 = weight(_text_:suchmaschine in 852) [ClassicSimilarity], result of:
          0.13238293 = score(doc=852,freq=2.0), product of:
            0.21191008 = queryWeight, product of:
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03747799 = queryNorm
            0.62471277 = fieldWeight in 852, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.078125 = fieldNorm(doc=852)
      0.16666667 = coord(1/6)
    
    Abstract
    Höchste Alarmstufe bei der weltgrößten Suchmaschine: Mit ChatGPT und künstlicher Intelligenz könnte eine neue Ära beginnen.
  13. Dietz, K.: en.wikipedia.org > 6 Mio. Artikel (2020) 0.02
    0.019381298 = product of:
      0.05814389 = sum of:
        0.049604155 = product of:
          0.14881246 = sum of:
            0.14881246 = weight(_text_:3a in 5669) [ClassicSimilarity], result of:
              0.14881246 = score(doc=5669,freq=2.0), product of:
                0.3177388 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.03747799 = queryNorm
                0.46834838 = fieldWeight in 5669, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5669)
          0.33333334 = coord(1/3)
        0.008539738 = product of:
          0.025619213 = sum of:
            0.025619213 = weight(_text_:29 in 5669) [ClassicSimilarity], result of:
              0.025619213 = score(doc=5669,freq=2.0), product of:
                0.13183585 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.03747799 = queryNorm
                0.19432661 = fieldWeight in 5669, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5669)
          0.33333334 = coord(1/3)
      0.33333334 = coord(2/6)
    
    Content
    "Die Englischsprachige Wikipedia verfügt jetzt über mehr als 6 Millionen Artikel. An zweiter Stelle kommt die deutschsprachige Wikipedia mit 2.3 Millionen Artikeln, an dritter Stelle steht die französischsprachige Wikipedia mit 2.1 Millionen Artikeln (via Researchbuzz: Firehose <https://rbfirehose.com/2020/01/24/techcrunch-wikipedia-now-has-more-than-6-million-articles-in-english/> und Techcrunch <https://techcrunch.com/2020/01/23/wikipedia-english-six-million-articles/?utm_source=feedburner&utm_medium=feed&utm_campaign=Feed%3A+Techcrunch+%28TechCrunch%29&guccounter=1&guce_referrer=aHR0cHM6Ly9yYmZpcmVob3NlLmNvbS8yMDIwLzAxLzI0L3RlY2hjcnVuY2gtd2lraXBlZGlhLW5vdy1oYXMtbW9yZS10aGFuLTYtbWlsbGlvbi1hcnRpY2xlcy1pbi1lbmdsaXNoLw&guce_referrer_sig=AQAAAK0zHfjdDZ_spFZBF_z-zDjtL5iWvuKDumFTzm4HvQzkUfE2pLXQzGS6FGB_y-VISdMEsUSvkNsg2U_NWQ4lwWSvOo3jvXo1I3GtgHpP8exukVxYAnn5mJspqX50VHIWFADHhs5AerkRn3hMRtf_R3F1qmEbo8EROZXp328HMC-o>). 250120 via digithek ch = #fineBlog s.a.: Angesichts der Veröffentlichung des 6-millionsten Artikels vergangene Woche in der englischsprachigen Wikipedia hat die Community-Zeitungsseite "Wikipedia Signpost" ein Moratorium bei der Veröffentlichung von Unternehmensartikeln gefordert. Das sei kein Vorwurf gegen die Wikimedia Foundation, aber die derzeitigen Maßnahmen, um die Enzyklopädie gegen missbräuchliches undeklariertes Paid Editing zu schützen, funktionierten ganz klar nicht. *"Da die ehrenamtlichen Autoren derzeit von Werbung in Gestalt von Wikipedia-Artikeln überwältigt werden, und da die WMF nicht in der Lage zu sein scheint, dem irgendetwas entgegenzusetzen, wäre der einzige gangbare Weg für die Autoren, fürs erste die Neuanlage von Artikeln über Unternehmen zu untersagen"*, schreibt der Benutzer Smallbones in seinem Editorial <https://en.wikipedia.org/wiki/Wikipedia:Wikipedia_Signpost/2020-01-27/From_the_editor> zur heutigen Ausgabe."
  14. Leisinger, C.: Sobald die Konkurrenten eine faire Chance haben, wird Google auf einen Schlag 20 Prozent seines Marktanteils verlieren (2020) 0.02
    0.017651059 = product of:
      0.105906345 = sum of:
        0.105906345 = weight(_text_:suchmaschine in 44) [ClassicSimilarity], result of:
          0.105906345 = score(doc=44,freq=2.0), product of:
            0.21191008 = queryWeight, product of:
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03747799 = queryNorm
            0.4997702 = fieldWeight in 44, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.0625 = fieldNorm(doc=44)
      0.16666667 = coord(1/6)
    
    Abstract
    Gabriel Weinberg macht Google und Co. mit seiner Suchmaschine DuckDuckGo immer mehr Konkurrenz. Im Gespräch erklärt er, wieso er auf den diskreten Umgang mit persönlichen Daten so viel Wert legt und warum er vehement für regulatorische Eingriffe ist.
  15. Daquino, M.: ¬A computational analysis of art historical linked data for assessing authoritativeness of attributions (2020) 0.02
    0.017133057 = product of:
      0.102798335 = sum of:
        0.102798335 = weight(_text_:ranking in 5916) [ClassicSimilarity], result of:
          0.102798335 = score(doc=5916,freq=4.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.5070964 = fieldWeight in 5916, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.046875 = fieldNorm(doc=5916)
      0.16666667 = coord(1/6)
    
    Abstract
    In this article a comparative analysis of art historical linked open data are presented. The result of the analysis is a conceptual framework of Information Quality (IQ) measures designed for validating contradictory sources of attribution on the basis of a documentary, evidence-based approach. The aim is to develop an ontology-based ranking model for recommending artwork attributions and support historians and catalogers' decision-making process. The conceptual framework was evaluated by means of a user study and the evaluation of a web application leveraging the aforementioned ranking model. The results of the survey demonstrate that the findings satisfy users' expectations and are potentially applicable to other types of information in the arts and humanities field.
  16. Lewandowski, D.: Suchmaschinen verstehen : 3. vollständig überarbeitete und erweiterte Aufl. (2021) 0.02
    0.0156014785 = product of:
      0.09360887 = sum of:
        0.09360887 = weight(_text_:suchmaschine in 4016) [ClassicSimilarity], result of:
          0.09360887 = score(doc=4016,freq=4.0), product of:
            0.21191008 = queryWeight, product of:
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03747799 = queryNorm
            0.44173864 = fieldWeight in 4016, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4016)
      0.16666667 = coord(1/6)
    
    RSWK
    Suchmaschine
    Subject
    Suchmaschine
  17. Irle, G.: Emotionen im Information Seeking (2023) 0.02
    0.015444675 = product of:
      0.09266805 = sum of:
        0.09266805 = weight(_text_:suchmaschine in 814) [ClassicSimilarity], result of:
          0.09266805 = score(doc=814,freq=2.0), product of:
            0.21191008 = queryWeight, product of:
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.03747799 = queryNorm
            0.43729892 = fieldWeight in 814, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.6542544 = idf(docFreq=420, maxDocs=44218)
              0.0546875 = fieldNorm(doc=814)
      0.16666667 = coord(1/6)
    
    Abstract
    In der Informationswissenschaft wurden Emotionen im Zuge des Affective Turn als zentraler Bestandteil des Forschungsinteresses anerkannt (vgl. Kapitel D 1 Information Behaviour, Informationsverhalten). Die Relevanz der Emotionsforschung ergibt sich durch ihre Bedeutung für eine wirksame und ganzheitliche Unterstützung der Suchenden. Die vier Bereiche, in denen die Emotionsforschung bei der Mensch-Maschine-Interaktion dienen kann, können auf das Information Seeking übertragen werden: 1. Systeme erkennen Affekt beim Suchenden, 2. Systeme passen ihre Funktionalität, wie zum Beispiel die Bedienelemente der Suchmaschine, an menschlichen Affekt an, 3. Systeme drücken Affekt aus, beispielsweise mit Avataren, 4. Menschlicher oder maschineller Affekt wird modelliert.
  18. Gao, R.; Ge, Y.; Sha, C.: FAIR: Fairness-aware information retrieval evaluation (2022) 0.01
    0.014277548 = product of:
      0.085665286 = sum of:
        0.085665286 = weight(_text_:ranking in 669) [ClassicSimilarity], result of:
          0.085665286 = score(doc=669,freq=4.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.42258036 = fieldWeight in 669, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.0390625 = fieldNorm(doc=669)
      0.16666667 = coord(1/6)
    
    Abstract
    With the emerging needs of creating fairness-aware solutions for search and recommendation systems, a daunting challenge exists of evaluating such solutions. While many of the traditional information retrieval (IR) metrics can capture the relevance, diversity, and novelty for the utility with respect to users, they are not suitable for inferring whether the presented results are fair from the perspective of responsible information exposure. On the other hand, existing fairness metrics do not account for user utility or do not measure it adequately. To address this problem, we propose a new metric called FAIR. By unifying standard IR metrics and fairness measures into an integrated metric, this metric offers a new perspective for evaluating fairness-aware ranking results. Based on this metric, we developed an effective ranking algorithm that jointly optimized user utility and fairness. The experimental results showed that our FAIR metric could highlight results with good user utility and fair information exposure. We showed how FAIR related to a set of existing utility and fairness metrics and demonstrated the effectiveness of our FAIR-based algorithm. We believe our work opens up a new direction of pursuing a metric for evaluating and implementing the FAIR systems.
  19. Haley, M.R.: ¬A simple paradigm for augmenting the Euclidean index to reflect journal impact and visibility (2020) 0.01
    0.0121149 = product of:
      0.0726894 = sum of:
        0.0726894 = weight(_text_:ranking in 5676) [ClassicSimilarity], result of:
          0.0726894 = score(doc=5676,freq=2.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.35857132 = fieldWeight in 5676, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.046875 = fieldNorm(doc=5676)
      0.16666667 = coord(1/6)
    
    Abstract
    This article offers an adjustment to the recently developed Euclidean Index (Perry and Reny, 2016). The proposed companion metric reflects the impact of the journal in which an article appears; the rationale for incorporating this information is to reflect higher costs of production and higher review standards, and to mitigate the heavily truncated citation counts that often arise in promotion, renewal, and tenure deliberations. Additionally, focusing jointly on citations and journal impact diversifies the assessment process, and can thereby help avoid misjudging scholars with modest citation counts in high-level journals. A combination of both metrics is also proposed, which nests each as a special case. The approach is demonstrated using a generic journal ranking metric, but can be adapted to most any stated or revealed preference measure of journal impact.
  20. Kang, X.; Wu, Y.; Ren, W.: Toward action comprehension for searching : mining actionable intents in query entities (2020) 0.01
    0.010095751 = product of:
      0.0605745 = sum of:
        0.0605745 = weight(_text_:ranking in 5613) [ClassicSimilarity], result of:
          0.0605745 = score(doc=5613,freq=2.0), product of:
            0.20271951 = queryWeight, product of:
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.03747799 = queryNorm
            0.29880944 = fieldWeight in 5613, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.4090285 = idf(docFreq=537, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5613)
      0.16666667 = coord(1/6)
    
    Abstract
    Understanding search engine users' intents has been a popular study in information retrieval, which directly affects the quality of retrieved information. One of the fundamental problems in this field is to find a connection between the entity in a query and the potential intents of the users, the latter of which would further reveal important information for facilitating the users' future actions. In this article, we present a novel research method for mining the actionable intents for search users, by generating a ranked list of the potentially most informative actions based on a massive pool of action samples. We compare different search strategies and their combinations for retrieving the action pool and develop three criteria for measuring the informativeness of the selected action samples, that is, the significance of an action sample within the pool, the representativeness of an action sample for the other candidate samples, and the diverseness of an action sample with respect to the selected actions. Our experiment, based on the Action Mining (AM) query entity data set from the Actionable Knowledge Graph (AKG) task at NTCIR-13, suggests that the proposed approach is effective in generating an informative and early-satisfying ranking of potential actions for search users.

Languages

  • e 137
  • d 76

Types

  • a 193
  • el 44
  • m 8
  • p 4
  • x 1
  • More… Less…