Search (85 results, page 4 of 5)

  • × theme_ss:"Semantisches Umfeld in Indexierung u. Retrieval"
  1. Mäkelä, E.; Hyvönen, E.; Saarela, S.; Vilfanen, K.: Application of ontology techniques to view-based semantic serach and browsing (2012) 0.00
    0.001153389 = product of:
      0.010380501 = sum of:
        0.010380501 = product of:
          0.020761002 = sum of:
            0.020761002 = weight(_text_:web in 3264) [ClassicSimilarity], result of:
              0.020761002 = score(doc=3264,freq=2.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.21634221 = fieldWeight in 3264, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3264)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    We scho how the beenfits of the view-based search method, developed within the information retrieval community, can be extended with ontology-based search, developed within the Semantic Web community, and with semantic recommendations. As a proof of the concept, we have implemented an ontology-and view-based search engine and recommendations system Ontogaotr for RDF(S) repositories. Ontogator is innovative in two ways. Firstly, the RDFS.based ontologies used for annotating metadata are used in the user interface to facilitate view-based information retrieval. The views provide the user with an overview of the repositorys contents and a vocabulary for expressing search queries. Secondlyy, a semantic browsing function is provided by a recommender system. This system enriches instance level metadata by ontologies and provides the user with links to semantically related relevant resources. The semantic linkage is specified in terms of logical rules. To illustrate and discuss the ideas, a deployed application of Ontogator to a photo repository of the Helsinki University Museum is presented.
  2. Efthimiadis, E.N.: User choices : a new yardstick for the evaluation of ranking algorithms for interactive query expansion (1995) 0.00
    0.001106661 = product of:
      0.009959949 = sum of:
        0.009959949 = product of:
          0.019919898 = sum of:
            0.019919898 = weight(_text_:22 in 5697) [ClassicSimilarity], result of:
              0.019919898 = score(doc=5697,freq=2.0), product of:
                0.10297151 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02940506 = queryNorm
                0.19345059 = fieldWeight in 5697, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5697)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Date
    22. 2.1996 13:14:10
  3. Song, D.; Bruza, P.D.: Towards context sensitive information inference (2003) 0.00
    0.001106661 = product of:
      0.009959949 = sum of:
        0.009959949 = product of:
          0.019919898 = sum of:
            0.019919898 = weight(_text_:22 in 1428) [ClassicSimilarity], result of:
              0.019919898 = score(doc=1428,freq=2.0), product of:
                0.10297151 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02940506 = queryNorm
                0.19345059 = fieldWeight in 1428, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1428)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Date
    22. 3.2003 19:35:46
  4. Hannech, A.: Système de recherche d'information étendue basé sur une projection multi-espaces (2018) 0.00
    0.0010874257 = product of:
      0.009786831 = sum of:
        0.009786831 = product of:
          0.019573662 = sum of:
            0.019573662 = weight(_text_:web in 4472) [ClassicSimilarity], result of:
              0.019573662 = score(doc=4472,freq=16.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.2039694 = fieldWeight in 4472, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.015625 = fieldNorm(doc=4472)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    Depuis son apparition au début des années 90, le World Wide Web (WWW ou Web) a offert un accès universel aux connaissances et le monde de l'information a été principalement témoin d'une grande révolution (la révolution numérique). Il est devenu rapidement très populaire, ce qui a fait de lui la plus grande et vaste base de données et de connaissances existantes grâce à la quantité et la diversité des données qu'il contient. Cependant, l'augmentation et l'évolution considérables de ces données soulèvent d'importants problèmes pour les utilisateurs notamment pour l'accès aux documents les plus pertinents à leurs requêtes de recherche. Afin de faire face à cette explosion exponentielle du volume de données et faciliter leur accès par les utilisateurs, différents modèles sont proposés par les systèmes de recherche d'information (SRIs) pour la représentation et la recherche des documents web. Les SRIs traditionnels utilisent, pour indexer et récupérer ces documents, des mots-clés simples qui ne sont pas sémantiquement liés. Cela engendre des limites en termes de la pertinence et de la facilité d'exploration des résultats. Pour surmonter ces limites, les techniques existantes enrichissent les documents en intégrant des mots-clés externes provenant de différentes sources. Cependant, ces systèmes souffrent encore de limitations qui sont liées aux techniques d'exploitation de ces sources d'enrichissement. Lorsque les différentes sources sont utilisées de telle sorte qu'elles ne peuvent être distinguées par le système, cela limite la flexibilité des modèles d'exploration qui peuvent être appliqués aux résultats de recherche retournés par ce système. Les utilisateurs se sentent alors perdus devant ces résultats, et se retrouvent dans l'obligation de les filtrer manuellement pour sélectionner l'information pertinente. S'ils veulent aller plus loin, ils doivent reformuler et cibler encore plus leurs requêtes de recherche jusqu'à parvenir aux documents qui répondent le mieux à leurs attentes. De cette façon, même si les systèmes parviennent à retrouver davantage des résultats pertinents, leur présentation reste problématique. Afin de cibler la recherche à des besoins d'information plus spécifiques de l'utilisateur et améliorer la pertinence et l'exploration de ses résultats de recherche, les SRIs avancés adoptent différentes techniques de personnalisation de données qui supposent que la recherche actuelle d'un utilisateur est directement liée à son profil et/ou à ses expériences de navigation/recherche antérieures. Cependant, cette hypothèse ne tient pas dans tous les cas, les besoins de l'utilisateur évoluent au fil du temps et peuvent s'éloigner de ses intérêts antérieurs stockés dans son profil.
    Dans d'autres cas, le profil de l'utilisateur peut être mal exploité pour extraire ou inférer ses nouveaux besoins en information. Ce problème est beaucoup plus accentué avec les requêtes ambigües. Lorsque plusieurs centres d'intérêt auxquels est liée une requête ambiguë sont identifiés dans le profil de l'utilisateur, le système se voit incapable de sélectionner les données pertinentes depuis ce profil pour répondre à la requête. Ceci a un impact direct sur la qualité des résultats fournis à cet utilisateur. Afin de remédier à quelques-unes de ces limitations, nous nous sommes intéressés dans ce cadre de cette thèse de recherche au développement de techniques destinées principalement à l'amélioration de la pertinence des résultats des SRIs actuels et à faciliter l'exploration de grandes collections de documents. Pour ce faire, nous proposons une solution basée sur un nouveau concept d'indexation et de recherche d'information appelé la projection multi-espaces. Cette proposition repose sur l'exploitation de différentes catégories d'information sémantiques et sociales qui permettent d'enrichir l'univers de représentation des documents et des requêtes de recherche en plusieurs dimensions d'interprétations. L'originalité de cette représentation est de pouvoir distinguer entre les différentes interprétations utilisées pour la description et la recherche des documents. Ceci donne une meilleure visibilité sur les résultats retournés et aide à apporter une meilleure flexibilité de recherche et d'exploration, en donnant à l'utilisateur la possibilité de naviguer une ou plusieurs vues de données qui l'intéressent le plus. En outre, les univers multidimensionnels de représentation proposés pour la description des documents et l'interprétation des requêtes de recherche aident à améliorer la pertinence des résultats de l'utilisateur en offrant une diversité de recherche/exploration qui aide à répondre à ses différents besoins et à ceux des autres différents utilisateurs. Cette étude exploite différents aspects liés à la recherche personnalisée et vise à résoudre les problèmes engendrés par l'évolution des besoins en information de l'utilisateur. Ainsi, lorsque le profil de cet utilisateur est utilisé par notre système, une technique est proposée et employée pour identifier les intérêts les plus représentatifs de ses besoins actuels dans son profil. Cette technique se base sur la combinaison de trois facteurs influents, notamment le facteur contextuel, fréquentiel et temporel des données. La capacité des utilisateurs à interagir, à échanger des idées et d'opinions, et à former des réseaux sociaux sur le Web, a amené les systèmes à s'intéresser aux types d'interactions de ces utilisateurs, au niveau d'interaction entre eux ainsi qu'à leurs rôles sociaux dans le système. Ces informations sociales sont abordées et intégrées dans ce travail de recherche. L'impact et la manière de leur intégration dans le processus de RI sont étudiés pour améliorer la pertinence des résultats.
    Since its appearance in the early 90's, the World Wide Web (WWW or Web) has provided universal access to knowledge and the world of information has been primarily witness to a great revolution (the digital revolution). It quickly became very popular, making it the largest and most comprehensive database and knowledge base thanks to the amount and diversity of data it contains. However, the considerable increase and evolution of these data raises important problems for users, in particular for accessing the documents most relevant to their search queries. In order to cope with this exponential explosion of data volume and facilitate their access by users, various models are offered by information retrieval systems (IRS) for the representation and retrieval of web documents. Traditional SRIs use simple keywords that are not semantically linked to index and retrieve these documents. This creates limitations in terms of the relevance and ease of exploration of results. To overcome these limitations, existing techniques enrich documents by integrating external keywords from different sources. However, these systems still suffer from limitations that are related to the exploitation techniques of these sources of enrichment. When the different sources are used so that they cannot be distinguished by the system, this limits the flexibility of the exploration models that can be applied to the results returned by this system. Users then feel lost to these results, and find themselves forced to filter them manually to select the relevant information. If they want to go further, they must reformulate and target their search queries even more until they reach the documents that best meet their expectations. In this way, even if the systems manage to find more relevant results, their presentation remains problematic. In order to target research to more user-specific information needs and improve the relevance and exploration of its research findings, advanced SRIs adopt different data personalization techniques that assume that current research of user is directly related to his profile and / or previous browsing / search experiences.
    However, this assumption does not hold in all cases, the needs of the user evolve over time and can move away from his previous interests stored in his profile. In other cases, the user's profile may be misused to extract or infer new information needs. This problem is much more accentuated with ambiguous queries. When multiple POIs linked to a search query are identified in the user's profile, the system is unable to select the relevant data from that profile to respond to that request. This has a direct impact on the quality of the results provided to this user. In order to overcome some of these limitations, in this research thesis, we have been interested in the development of techniques aimed mainly at improving the relevance of the results of current SRIs and facilitating the exploration of major collections of documents. To do this, we propose a solution based on a new concept and model of indexing and information retrieval called multi-spaces projection. This proposal is based on the exploitation of different categories of semantic and social information that enrich the universe of document representation and search queries in several dimensions of interpretations. The originality of this representation is to be able to distinguish between the different interpretations used for the description and the search for documents. This gives a better visibility on the results returned and helps to provide a greater flexibility of search and exploration, giving the user the ability to navigate one or more views of data that interest him the most. In addition, the proposed multidimensional representation universes for document description and search query interpretation help to improve the relevance of the user's results by providing a diversity of research / exploration that helps meet his diverse needs and those of other different users. This study exploits different aspects that are related to the personalized search and aims to solve the problems caused by the evolution of the information needs of the user. Thus, when the profile of this user is used by our system, a technique is proposed and used to identify the interests most representative of his current needs in his profile. This technique is based on the combination of three influential factors, including the contextual, frequency and temporal factor of the data. The ability of users to interact, exchange ideas and opinions, and form social networks on the Web, has led systems to focus on the types of interactions these users have at the level of interaction between them as well as their social roles in the system. This social information is discussed and integrated into this research work. The impact and how they are integrated into the IR process are studied to improve the relevance of the results.
  5. Case, D.O.: Looking for information : a survey on research on information seeking, needs, and behavior (2002) 0.00
    9.988644E-4 = product of:
      0.008989779 = sum of:
        0.008989779 = product of:
          0.017979559 = sum of:
            0.017979559 = weight(_text_:web in 1270) [ClassicSimilarity], result of:
              0.017979559 = score(doc=1270,freq=6.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.18735787 = fieldWeight in 1270, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1270)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Footnote
    Rez. in: JASIST 54(2003) no.7, S.695-697 (R. Savolainen): "Donald O. Case has written an ambitious book to create an overall picture of the major approaches to information needs and seeking (INS) studies. The aim to write an extensive review is reflected in the list of references containing about 700 items. The high ambitions are explained an p. 14, where Case states that he is aiming at a multidisciplinary understanding of the concept of information seeking. In the Preface, the author characterizes his book as an introduction to the topic for students at the graduate level, as well as as a review and handbook for scholars engagged in information behavior research. In my view, Looking for Information is particularly welcome as an academic textbook because the field of INS studies suffers from the lack of monographs. Along with the continuous growth of the number of journal articles and conference papers, there is a genuine need for a book that picks up the numerous pieces and puts them together. The use of the study as a textbook is facilitated by clearly delineated sections an major themes and the wealth of concrete examples of information seeking in everyday contexts. The book is lucidly written and it is accessible to novice readers, too. At first glance, the idea of providing a comprehensive review of INS studies may seem a mission impossible because the current number of articles, papers, and other contributions in this field is nearing the 10,000 range (p. 224). Donald Case is not alone in the task of coming to grips with an increasing number of studies; similar problems have been faced by those writing INS-related chapters for the Annual Review of Information Science and Technology (ARIST). Case has solved the problem of "too many publications to be reviewed" by concentrating an the INS literature published during the last two decades. Secondly, studies an library use and information retrieval are discussed only to a limited extent. In addition, Case is highly selective as to studies focusing an the use of specific sources and channels such as WWW. These delineations are reasonable, even though they beg some questions. First, how should one draw the line between studies an information seeking and information retrieval? Case does not discuss this question in greater detail, although in recent years, the overlapping areas of information seeking and retrieval studies have been broadened, along with the growing importance of WWW in information seeking/retrieval. Secondly, how can one define the concept of information searching (or, more specifically, Internet or Web searching) in relation to information seeking and information retrieval? In the field of Web searching studies, there is an increasing number of contributions that are of direct relevance to information-seeking studies. Clearly, the advent of the Internet, particularly, the Web, has blurred the previous lines between INS and IR literature, making them less clear cut. The book consists of five main sections, and comprises 13 chapters. There is an Appendix serving the needs of an INS textbook (questions for discussion and application). The structure of the book is meticulously planned and, as a whole, it offers a sufficiently balanced contribution to theoretical, methodological, and empirical issues of INS. The title, Looking for Information: A Survey of Research an Information Seeking, Needs, and Behavior aptly describes the main substance of the book. . . . It is easy to agree with Case about the significance of the problem of specialization and fragmentation. This problem seems to be concomitant with the broadening field of INS research. In itself, Case's book can be interpreted as a struggle against this fragmentation. His book suggests that this struggle is not hopeless and that it is still possible to draw an overall picture of the evolving research field. The major pieces of the puzzle were found and the book will provide a useful overview of INS studies for many years."
  6. Tudhope, D.; Blocks, D.; Cunliffe, D.; Binding, C.: Query expansion via conceptual distance in thesaurus indexed collections (2006) 0.00
    9.611576E-4 = product of:
      0.008650418 = sum of:
        0.008650418 = product of:
          0.017300837 = sum of:
            0.017300837 = weight(_text_:web in 2215) [ClassicSimilarity], result of:
              0.017300837 = score(doc=2215,freq=2.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.18028519 = fieldWeight in 2215, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2215)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    Purpose - The purpose of this paper is to explore query expansion via conceptual distance in thesaurus indexed collections Design/methodology/approach - An extract of the National Museum of Science and Industry's collections database, indexed with the Getty Art and Architecture Thesaurus (AAT), was the dataset for the research. The system architecture and algorithms for semantic closeness and the matching function are outlined. Standalone and web interfaces are described and formative qualitative user studies are discussed. One user session is discussed in detail, together with a scenario based on a related public inquiry. Findings are set in context of the literature on thesaurus-based query expansion. This paper discusses the potential of query expansion techniques using the semantic relationships in a faceted thesaurus. Findings - Thesaurus-assisted retrieval systems have potential for multi-concept descriptors, permitting very precise queries and indexing. However, indexer and searcher may differ in terminology judgments and there may not be any exactly matching results. The integration of semantic closeness in the matching function permits ranked results for multi-concept queries in thesaurus-indexed applications. An in-memory representation of the thesaurus semantic network allows a combination of automatic and interactive control of expansion and control of expansion on individual query terms. Originality/value - The application of semantic expansion to browsing may be useful in interface options where thesaurus structure is hidden.
  7. Rädler, K.: In Bibliothekskatalogen "googlen" : Integration von Inhaltsverzeichnissen, Volltexten und WEB-Ressourcen in Bibliothekskataloge (2004) 0.00
    9.611576E-4 = product of:
      0.008650418 = sum of:
        0.008650418 = product of:
          0.017300837 = sum of:
            0.017300837 = weight(_text_:web in 2432) [ClassicSimilarity], result of:
              0.017300837 = score(doc=2432,freq=2.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.18028519 = fieldWeight in 2432, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2432)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
  8. Kruschwitz, U.; AI-Bakour, H.: Users want more sophisticated search assistants : results of a task-based evaluation (2005) 0.00
    9.611576E-4 = product of:
      0.008650418 = sum of:
        0.008650418 = product of:
          0.017300837 = sum of:
            0.017300837 = weight(_text_:web in 4575) [ClassicSimilarity], result of:
              0.017300837 = score(doc=4575,freq=2.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.18028519 = fieldWeight in 4575, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4575)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    The Web provides a massive knowledge source, as do intranets and other electronic document collections. However, much of that knowledge is encoded implicitly and cannot be applied directly without processing into some more appropriate structures. Searching, browsing, question answering, for example, could all benefit from domain-specific knowledge contained in the documents, and in applications such as simple search we do not actually need very "deep" knowledge structures such as ontologies, but we can get a long way with a model of the domain that consists of term hierarchies. We combine domain knowledge automatically acquired by exploiting the documents' markup structure with knowledge extracted an the fly to assist a user with ad hoc search requests. Such a search system can suggest query modification options derived from the actual data and thus guide a user through the space of documents. This article gives a detailed account of a task-based evaluation that compares a search system that uses the outlined domain knowledge with a standard search system. We found that users do use the query modification suggestions proposed by the system. The main conclusion we can draw from this evaluation, however, is that users prefer a system that can suggest query modifications over a standard search engine, which simply presents a ranked list of documents. Most interestingly, we observe this user preference despite the fact that the baseline system even performs slightly better under certain criteria.
  9. Weichselgartner, E.: ZPID bindet Thesaurus in Retrievaloberfläche ein (2006) 0.00
    9.611576E-4 = product of:
      0.008650418 = sum of:
        0.008650418 = product of:
          0.017300837 = sum of:
            0.017300837 = weight(_text_:web in 5962) [ClassicSimilarity], result of:
              0.017300837 = score(doc=5962,freq=2.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.18028519 = fieldWeight in 5962, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5962)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    Seit 3. Juli 2006 stellt das ZPID eine verbesserte Suchoberfläche für die Recherche in der bibliographischen Psychologie-Datenbank PSYNDEX zur Verfügung. Hauptmerkmal der neuen Version 1.1 des 'ZPID-Retrieval für PSYNDEX' ist die Einbindung von 'PSYNDEX Terms', dem kontrollierten Wortschatz der psychologischen Fachsprache. PSYNDEX Terms basiert auf dem 'Thesaurus of Psychological Index Terms' der American Psychological Association (APA) und enthält im Moment über 5.400 Deskriptoren. Zu jedem Deskriptor werden ggf. Oberbegriffe, Unterbegriffe und verwandte Begriffe angezeigt. Wer die Suchoberfläche nutzt, kann entweder im Thesaurus blättern oder gezielt nach Thesaurusbegriffen suchen. Kommt der eigene frei gewählte Suchbegriff nicht im Thesaurus vor, macht das System selbsttätig Vorschläge für passende Thesaurusbegriffe. DerThesaurus ist komplett zweisprachig (deutsch/englisch) implementiert, sodass er auch als Übersetzungshilfe dient. Weitere Verbesserungen der Suchoberfläche betreffen die Darstellbarkeit in unterschiedlichen Web-Browsern mit dem Ziel der Barrierefreiheit, die Erweiterung der OnlineHilfe mit Beispielen für erfolgreiche Suchstrategien, die Möglichkeit, zu speziellen Themen vertiefte Informationen abzurufen (den Anfang machen psychologische Behandlungsprogramme) und die Bereitstellung eines Export-Filters für EndNote. Zielgruppe des ZPID-Retrieval sind Einzelpersonen, die keinen institutionellen PSYNDEX-Zugang, z.B. am Campus einer Universität, nutzen können. Sie können das kostenpflichtige Retrieval direkt online erwerben und werden binnen weniger Minuten freigeschaltet. Kunden mit existierendem Vertrag kommen automatisch in den Genuss der verbesserten Suchoberfläche.
  10. Melucci, M.: Contextual search : a computational framework (2012) 0.00
    9.611576E-4 = product of:
      0.008650418 = sum of:
        0.008650418 = product of:
          0.017300837 = sum of:
            0.017300837 = weight(_text_:web in 4913) [ClassicSimilarity], result of:
              0.017300837 = score(doc=4913,freq=2.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.18028519 = fieldWeight in 4913, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4913)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    The growing availability of data in electronic form, the expansion of the World Wide Web and the accessibility of computational methods for large-scale data processing have allowed researchers in Information Retrieval (IR) to design systems which can effectively and efficiently constrain search within the boundaries given by context, thus transforming classical search into contextual search. Contextual Search: A Computational Framework introduces contextual search within a computational framework based on contextual variables, contextual factors and statistical models. It describes how statistical models can process contextual variables to infer the contextual factors underlying the current search context. It also provides background to the subject by: placing it among other surveys on relevance, interaction, context, and behaviour; providing a description of the contextual variables used for implementing the statistical models which represent and predict relevance and contextual factors; and providing an overview of the evaluation methodologies and findings relevant to this subject. Contextual Search: A Computational Framework is a highly recommended read, both for beginners who are embarking on research in this area and as a useful reference for established IR researchers.
  11. Xamena, E.; Brignole, N.B.; Maguitman, A.G.: ¬A study of relevance propagation in large topic ontologies (2013) 0.00
    9.611576E-4 = product of:
      0.008650418 = sum of:
        0.008650418 = product of:
          0.017300837 = sum of:
            0.017300837 = weight(_text_:web in 1105) [ClassicSimilarity], result of:
              0.017300837 = score(doc=1105,freq=2.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.18028519 = fieldWeight in 1105, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1105)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    Topic ontologies or web directories consist of large collections of links to websites, arranged by topic in different categories. The structure of these ontologies is typically not flat because there are hierarchical and nonhierarchical relationships among topics. As a consequence, websites classified under a certain topic may be relevant to other topics. Although some of these relevance relations are explicit, most of them must be discovered by an analysis of the structure of the ontologies. This article proposes a family of models of relevance propagation in topic ontologies. An efficient computational framework is described and used to compute nine different models for a portion of the Open Directory Project graph consisting of more than half a million nodes and approximately 1.5 million edges of different types. After performing a quantitative analysis, a user study was carried out to compare the most promising models. It was found that some general difficulties rule out the possibility of defining flawless models of relevance propagation that only take into account structural aspects of an ontology. However, there is a clear indication that including transitive relations induced by the nonhierarchical components of the ontology results in relevance propagation models that are superior to more basic approaches.
  12. Symonds, M.; Bruza, P.; Zuccon, G.; Koopman, B.; Sitbon, L.; Turner, I.: Automatic query expansion : a structural linguistic perspective (2014) 0.00
    9.611576E-4 = product of:
      0.008650418 = sum of:
        0.008650418 = product of:
          0.017300837 = sum of:
            0.017300837 = weight(_text_:web in 1338) [ClassicSimilarity], result of:
              0.017300837 = score(doc=1338,freq=2.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.18028519 = fieldWeight in 1338, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1338)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    A user's query is considered to be an imprecise description of their information need. Automatic query expansion is the process of reformulating the original query with the goal of improving retrieval effectiveness. Many successful query expansion techniques model syntagmatic associations that infer two terms co-occur more often than by chance in natural language. However, structural linguistics relies on both syntagmatic and paradigmatic associations to deduce the meaning of a word. Given the success of dependency-based approaches to query expansion and the reliance on word meanings in the query formulation process, we argue that modeling both syntagmatic and paradigmatic information in the query expansion process improves retrieval effectiveness. This article develops and evaluates a new query expansion technique that is based on a formal, corpus-based model of word meaning that models syntagmatic and paradigmatic associations. We demonstrate that when sufficient statistical information exists, as in the case of longer queries, including paradigmatic information alone provides significant improvements in retrieval effectiveness across a wide variety of data sets. More generally, when our new query expansion approach is applied to large-scale web retrieval it demonstrates significant improvements in retrieval effectiveness over a strong baseline system, based on a commercial search engine.
  13. Cai, F.; Rijke, M. de: Learning from homologous queries and semantically related terms for query auto completion (2016) 0.00
    9.611576E-4 = product of:
      0.008650418 = sum of:
        0.008650418 = product of:
          0.017300837 = sum of:
            0.017300837 = weight(_text_:web in 2971) [ClassicSimilarity], result of:
              0.017300837 = score(doc=2971,freq=2.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.18028519 = fieldWeight in 2971, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2971)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    Query auto completion (QAC) models recommend possible queries to web search users when they start typing a query prefix. Most of today's QAC models rank candidate queries by popularity (i.e., frequency), and in doing so they tend to follow a strict query matching policy when counting the queries. That is, they ignore the contributions from so-called homologous queries, queries with the same terms but ordered differently or queries that expand the original query. Importantly, homologous queries often express a remarkably similar search intent. Moreover, today's QAC approaches often ignore semantically related terms. We argue that users are prone to combine semantically related terms when generating queries. We propose a learning to rank-based QAC approach, where, for the first time, features derived from homologous queries and semantically related terms are introduced. In particular, we consider: (i) the observed and predicted popularity of homologous queries for a query candidate; and (ii) the semantic relatedness of pairs of terms inside a query and pairs of queries inside a session. We quantify the improvement of the proposed new features using two large-scale real-world query logs and show that the mean reciprocal rank and the success rate can be improved by up to 9% over state-of-the-art QAC models.
  14. Ma, N.; Zheng, H.T.; Xiao, X.: ¬An ontology-based latent semantic indexing approach using long short-term memory networks (2017) 0.00
    9.611576E-4 = product of:
      0.008650418 = sum of:
        0.008650418 = product of:
          0.017300837 = sum of:
            0.017300837 = weight(_text_:web in 3810) [ClassicSimilarity], result of:
              0.017300837 = score(doc=3810,freq=2.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.18028519 = fieldWeight in 3810, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3810)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Source
    Web and Big Data: First International Joint Conference, APWeb-WAIM 2017, Beijing, China, July 7-9, 2017, Proceedings, Part I. Eds.: L. Chen et al
  15. Ferreira, R.S.; Graça Pimentel, M. de; Cristo, M.: ¬A wikification prediction model based on the combination of latent, dyadic, and monadic features (2018) 0.00
    9.611576E-4 = product of:
      0.008650418 = sum of:
        0.008650418 = product of:
          0.017300837 = sum of:
            0.017300837 = weight(_text_:web in 4119) [ClassicSimilarity], result of:
              0.017300837 = score(doc=4119,freq=2.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.18028519 = fieldWeight in 4119, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4119)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    Considering repositories of web documents that are semantically linked and created in a collaborative fashion, as in the case of Wikipedia, a key problem faced by content providers is the placement of links in the articles. These links must support user navigation and provide a deeper semantic interpretation of the content. Current wikification methods exploit machine learning techniques to capture characteristics of the concepts and its associations. In previous work, we proposed a preliminary prediction model combining traditional predictors with a latent component which captures the concept graph topology by means of matrix factorization. In this work, we provide a detailed description of our method and a deeper comparison with a state-of-the-art wikification method using a sample of Wikipedia and report a gain up to 13% in F1 score. We also provide a comprehensive analysis of the model performance showing the importance of the latent predictor component and the attributes derived from the associations between the concepts. Moreover, we include an analysis that allows us to conclude that the model is resilient to ambiguity without including a disambiguation phase. We finally report the positive impact of selecting training samples from specific content quality classes.
  16. Bradford, R.B.: Relationship discovery in large text collections using Latent Semantic Indexing (2006) 0.00
    8.853288E-4 = product of:
      0.007967959 = sum of:
        0.007967959 = product of:
          0.015935918 = sum of:
            0.015935918 = weight(_text_:22 in 1163) [ClassicSimilarity], result of:
              0.015935918 = score(doc=1163,freq=2.0), product of:
                0.10297151 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02940506 = queryNorm
                0.15476047 = fieldWeight in 1163, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1163)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Source
    Proceedings of the Fourth Workshop on Link Analysis, Counterterrorism, and Security, SIAM Data Mining Conference, Bethesda, MD, 20-22 April, 2006. [http://www.siam.org/meetings/sdm06/workproceed/Link%20Analysis/15.pdf]
  17. Gillitzer, B.: Yewno (2017) 0.00
    8.853288E-4 = product of:
      0.007967959 = sum of:
        0.007967959 = product of:
          0.015935918 = sum of:
            0.015935918 = weight(_text_:22 in 3447) [ClassicSimilarity], result of:
              0.015935918 = score(doc=3447,freq=2.0), product of:
                0.10297151 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02940506 = queryNorm
                0.15476047 = fieldWeight in 3447, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3447)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Date
    22. 2.2017 10:16:49
  18. Thenmalar, S.; Geetha, T.V.: Enhanced ontology-based indexing and searching (2014) 0.00
    7.746627E-4 = product of:
      0.006971964 = sum of:
        0.006971964 = product of:
          0.013943928 = sum of:
            0.013943928 = weight(_text_:22 in 1633) [ClassicSimilarity], result of:
              0.013943928 = score(doc=1633,freq=2.0), product of:
                0.10297151 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02940506 = queryNorm
                0.1354154 = fieldWeight in 1633, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1633)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Date
    20. 1.2015 18:30:22
  19. Efthimiadis, E.N.: Interactive query expansion : a user-based evaluation in a relevance feedback environment (2000) 0.00
    7.6892605E-4 = product of:
      0.0069203344 = sum of:
        0.0069203344 = product of:
          0.013840669 = sum of:
            0.013840669 = weight(_text_:web in 5701) [ClassicSimilarity], result of:
              0.013840669 = score(doc=5701,freq=2.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.14422815 = fieldWeight in 5701, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.03125 = fieldNorm(doc=5701)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    A user-centered investigation of interactive query expansion within the context of a relevance feedback system is presented in this article. Data were collected from 25 searches using the INSPEC database. The data collection mechanisms included questionnaires, transaction logs, and relevance evaluations. The results discuss issues that relate to query expansion, retrieval effectiveness, the correspondence of the on-line-to-off-line relevance judgments, and the selection of terms for query expansion by users (interactive query expansion). The main conclusions drawn from the results of the study are that: (1) one-third of the terms presented to users in a list of candidate terms for query expansion was identified by the users as potentially useful for query expansion. (2) These terms were mainly judged as either variant expressions (synonyms) or alternative (related) terms to the initial query terms. However, a substantial portion of the selected terms were identified as representing new ideas. (3) The relationships identified between the five best terms selected by the users for query expansion and the initial query terms were that: (a) 34% of the query expansion terms have no relationship or other type of correspondence with a query term; (b) 66% of the remaining query expansion terms have a relationship to the query terms. These relationships were: narrower term (46%), broader term (3%), related term (17%). (4) The results provide evidence for the effectiveness of interactive query expansion. The initial search produced on average three highly relevant documents; the query expansion search produced on average nine further highly relevant documents. The conclusions highlight the need for more research on: interactive query expansion, the comparative evaluation of automatic vs. interactive query expansion, the study of weighted Webbased or Web-accessible retrieval systems in operational environments, and for user studies in searching ranked retrieval systems in general
  20. Brezillon, P.; Saker, I.: Modeling context in information seeking (1999) 0.00
    7.6892605E-4 = product of:
      0.0069203344 = sum of:
        0.0069203344 = product of:
          0.013840669 = sum of:
            0.013840669 = weight(_text_:web in 276) [ClassicSimilarity], result of:
              0.013840669 = score(doc=276,freq=2.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.14422815 = fieldWeight in 276, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.03125 = fieldNorm(doc=276)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    Context plays an important role in a number of domains where reasoning intervenes as in understanding, interpretation, diagnosis, etc. The reason is that reasoning activities heavily rely on a background (or experience) that is generally not made explicit and that gives a contextual dimension to knowledge. On the Web in December 1996, AItaVista gave more than 710000 pages containing the word context, when concept gives only 639000 references. A clear definition of this word stays to be found. There are several formal definitions of this concept (references are given in Brézillon, 1996): a set of preferences and/or beliefs, an infinite and only partially known collection of assumptions, a list of attributes, the product of an interpretation, possible worlds, assumptions under which a statement is true or false. One faces the same situation at the programming level: a collection of context schemas; a path in information retrieval; slots in object-oriented languages; a special, buffer-like data structure; a window on the screen, buttons which are functional customisable and shareable; an interpreter which controls the system's activity; the characteristics of the situation and the goals of the knowledge use; or entities (things or events) related in a certain way that permits to listen what is said and what is not said. Context is often assimilated at a set of restrictions (e.g., preconditions) that limit access to parts of the applications. The first works considering context explicitly are in Natural Language. Researchers in this domain focus on the linguistic context, sometimes associated with other types of contexts as: semantic context, cognitive context, physical and perceptual context, and social context (Bunt, 1997).

Years

Languages

  • e 66
  • d 18
  • f 1
  • More… Less…

Types

  • a 73
  • el 11
  • m 8
  • r 1
  • s 1
  • x 1
  • More… Less…