Search (5059 results, page 2 of 253)

  1. Hofstede, A.H.M. ter; Proper, H.A.; Van der Weide, T.P.: Exploiting fact verbalisation in conceptual information modelling (1997) 0.11
    0.11080296 = product of:
      0.16620444 = sum of:
        0.14280158 = weight(_text_:query in 2908) [ClassicSimilarity], result of:
          0.14280158 = score(doc=2908,freq=6.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.62258047 = fieldWeight in 2908, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2908)
        0.023402857 = product of:
          0.046805713 = sum of:
            0.046805713 = weight(_text_:22 in 2908) [ClassicSimilarity], result of:
              0.046805713 = score(doc=2908,freq=2.0), product of:
                0.1728227 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049352113 = queryNorm
                0.2708308 = fieldWeight in 2908, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2908)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Focuses on the information modelling side of conceptual modelling. Deals with the exploitation of fact verbalisations after finishing the actual information system. Verbalisations are used as input for the design of the so-called information model. Exploits these verbalisation in 4 directions: considers their use for a conceptual query language, the verbalisation of instances, the description of the contents of a database and for the verbalisation of queries in a computer supported query environment. Provides an example session with an envisioned tool for end user query formulations that exploits the verbalisation
    Source
    Information systems. 22(1997) nos.5/6, S.349-385
  2. Pasicznyuk, R.W.: Searching for the information on the Net : new wine into new wine skins (1995) 0.11
    0.1081783 = product of:
      0.16226745 = sum of:
        0.09422461 = weight(_text_:query in 5785) [ClassicSimilarity], result of:
          0.09422461 = score(doc=5785,freq=2.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.41079655 = fieldWeight in 5785, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.0625 = fieldNorm(doc=5785)
        0.06804284 = product of:
          0.13608567 = sum of:
            0.13608567 = weight(_text_:page in 5785) [ClassicSimilarity], result of:
              0.13608567 = score(doc=5785,freq=2.0), product of:
                0.27565226 = queryWeight, product of:
                  5.5854197 = idf(docFreq=450, maxDocs=44218)
                  0.049352113 = queryNorm
                0.49368602 = fieldWeight in 5785, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.5854197 = idf(docFreq=450, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5785)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Provides a glossary of Internet search terms. Outlines a number of network retrieval tools and directories: Netscape's Internet search page, W3 search engines, Lycos, WebCrawler, InfoSeek, Yahoo, and CERN's Net Directory. Gices an example of how the Internet can be used to answer a reference query and the types of materials that can be retrieved
  3. Efthimiadis, E.N.: User choices : a new yardstick for the evaluation of ranking algorithms for interactive query expansion (1995) 0.11
    0.10731181 = product of:
      0.16096771 = sum of:
        0.14425138 = weight(_text_:query in 5697) [ClassicSimilarity], result of:
          0.14425138 = score(doc=5697,freq=12.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.6289012 = fieldWeight in 5697, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5697)
        0.016716326 = product of:
          0.03343265 = sum of:
            0.03343265 = weight(_text_:22 in 5697) [ClassicSimilarity], result of:
              0.03343265 = score(doc=5697,freq=2.0), product of:
                0.1728227 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049352113 = queryNorm
                0.19345059 = fieldWeight in 5697, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5697)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The performance of 8 ranking algorithms was evaluated with respect to their effectiveness in ranking terms for query expansion. The evaluation was conducted within an investigation of interactive query expansion and relevance feedback in a real operational environment. Focuses on the identification of algorithms that most effectively take cognizance of user preferences. user choices (i.e. the terms selected by the searchers for the query expansion search) provided the yardstick for the evaluation of the 8 ranking algorithms. This methodology introduces a user oriented approach in evaluating ranking algorithms for query expansion in contrast to the standard, system oriented approaches. Similarities in the performance of the 8 algorithms and the ways these algorithms rank terms were the main focus of this evaluation. The findings demonstrate that the r-lohi, wpq, enim, and porter algorithms have similar performance in bringing good terms to the top of a ranked list of terms for query expansion. However, further evaluation of the algorithms in different (e.g. full text) environments is needed before these results can be generalized beyond the context of the present study
    Date
    22. 2.1996 13:14:10
  4. Larkey, L.S.; Connell, M.E.: Structured queries, language modelling, and relevance modelling in cross-language information retrieval (2005) 0.11
    0.10731181 = product of:
      0.16096771 = sum of:
        0.14425138 = weight(_text_:query in 1022) [ClassicSimilarity], result of:
          0.14425138 = score(doc=1022,freq=12.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.6289012 = fieldWeight in 1022, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1022)
        0.016716326 = product of:
          0.03343265 = sum of:
            0.03343265 = weight(_text_:22 in 1022) [ClassicSimilarity], result of:
              0.03343265 = score(doc=1022,freq=2.0), product of:
                0.1728227 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049352113 = queryNorm
                0.19345059 = fieldWeight in 1022, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1022)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Two probabilistic approaches to cross-lingual retrieval are in wide use today, those based on probabilistic models of relevance, as exemplified by INQUERY, and those based on language modeling. INQUERY, as a query net model, allows the easy incorporation of query operators, including a synonym operator, which has proven to be extremely useful in cross-language information retrieval (CLIR), in an approach often called structured query translation. In contrast, language models incorporate translation probabilities into a unified framework. We compare the two approaches on Arabic and Spanish data sets, using two kinds of bilingual dictionaries--one derived from a conventional dictionary, and one derived from a parallel corpus. We find that structured query processing gives slightly better results when queries are not expanded. On the other hand, when queries are expanded, language modeling gives better results, but only when using a probabilistic dictionary derived from a parallel corpus. We pursue two additional issues inherent in the comparison of structured query processing with language modeling. The first concerns query expansion, and the second is the role of translation probabilities. We compare conventional expansion techniques (pseudo-relevance feedback) with relevance modeling, a new IR approach which fits into the formal framework of language modeling. We find that relevance modeling and pseudo-relevance feedback achieve comparable levels of retrieval and that good translation probabilities confer a small but significant advantage.
    Date
    26.12.2007 20:22:11
  5. Niu, X.; Kelly, D.: ¬The use of query suggestions during information search (2014) 0.11
    0.10731181 = product of:
      0.16096771 = sum of:
        0.14425138 = weight(_text_:query in 2702) [ClassicSimilarity], result of:
          0.14425138 = score(doc=2702,freq=12.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.6289012 = fieldWeight in 2702, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2702)
        0.016716326 = product of:
          0.03343265 = sum of:
            0.03343265 = weight(_text_:22 in 2702) [ClassicSimilarity], result of:
              0.03343265 = score(doc=2702,freq=2.0), product of:
                0.1728227 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049352113 = queryNorm
                0.19345059 = fieldWeight in 2702, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2702)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Query suggestion is a common feature of many information search systems. While much research has been conducted about how to generate suggestions, fewer studies have been conducted about how people interact with and use suggestions. The purpose of this paper is to investigate how and when people integrate query suggestions into their searches and the outcome of this usage. The paper further investigates the relationships between search expertise, topic difficulty, and temporal segment of the search and query suggestion usage. A secondary analysis of data was conducted using data collected in a previous controlled laboratory study. In this previous study, 23 undergraduate research participants used an experimental search system with query suggestions to conduct four topic searches. Results showed that participants integrated the suggestions into their searching fairly quickly and that participants with less search expertise used more suggestions and saved more documents. Participants also used more suggestions towards the end of their searches and when searching for more difficult topics. These results show that query suggestion can provide support in situations where people have less search expertise, greater difficulty searching and at specific times during the search.
    Date
    25. 1.2016 18:43:22
  6. Dimitrova, N.; Golshani, F.: Motion recovery for video content classification (1995) 0.11
    0.106666565 = product of:
      0.15999985 = sum of:
        0.13325372 = weight(_text_:query in 3834) [ClassicSimilarity], result of:
          0.13325372 = score(doc=3834,freq=4.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.5809541 = fieldWeight in 3834, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.0625 = fieldNorm(doc=3834)
        0.026746122 = product of:
          0.053492244 = sum of:
            0.053492244 = weight(_text_:22 in 3834) [ClassicSimilarity], result of:
              0.053492244 = score(doc=3834,freq=2.0), product of:
                0.1728227 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049352113 = queryNorm
                0.30952093 = fieldWeight in 3834, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3834)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Discusses the analysis of video for the classification of images in order to develop a video database. Covers compression; motion recovery in digital video; low-level motion extraction; single macroblock tracing; intermediate-level motion analysis; high-level motion analysis; spatiotemporal hierarchical representation; information filtering and digital video; content filtering opertaors; the query language; querying video contents; an architecture for video classification and retrieval; the visual query language VEVA; and implementation of macroblock tracing
    Date
    8. 4.1996 9:22:36
  7. Lee, D.; Srivastava, S.; Vista, D.: Generating advanced query interfaces (1998) 0.11
    0.106666565 = product of:
      0.15999985 = sum of:
        0.13325372 = weight(_text_:query in 3607) [ClassicSimilarity], result of:
          0.13325372 = score(doc=3607,freq=4.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.5809541 = fieldWeight in 3607, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.0625 = fieldNorm(doc=3607)
        0.026746122 = product of:
          0.053492244 = sum of:
            0.053492244 = weight(_text_:22 in 3607) [ClassicSimilarity], result of:
              0.053492244 = score(doc=3607,freq=2.0), product of:
                0.1728227 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049352113 = queryNorm
                0.30952093 = fieldWeight in 3607, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3607)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Describes the design and implementation of an interface generator for constructing advanced visual query WWW interfaces that allow the specification of complex queries. The generated inerfaces share a consistent look and feel. The tool accepts as input a high-level specification of the interface and produces as output its implementation
    Date
    1. 8.1996 22:08:06
  8. Haslhofer, B.: Uniform SPARQL access to interlinked (digital library) sources (2007) 0.11
    0.106666565 = product of:
      0.15999985 = sum of:
        0.13325372 = weight(_text_:query in 541) [ClassicSimilarity], result of:
          0.13325372 = score(doc=541,freq=4.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.5809541 = fieldWeight in 541, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.0625 = fieldNorm(doc=541)
        0.026746122 = product of:
          0.053492244 = sum of:
            0.053492244 = weight(_text_:22 in 541) [ClassicSimilarity], result of:
              0.053492244 = score(doc=541,freq=2.0), product of:
                0.1728227 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049352113 = queryNorm
                0.30952093 = fieldWeight in 541, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=541)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    In this presentation, we therefore focus on a solution for providing uniform access to Digital Libraries and other online services. In order to enable uniform query access to heterogeneous sources, we must provide metadata interoperability in a way that a query language - in this case SPARQL - can cope with the incompatibility of the metadata in various sources without changing their already existing information models.
    Date
    26.12.2011 13:22:46
  9. Lewandowski, D.: Query understanding (2011) 0.11
    0.106666565 = product of:
      0.15999985 = sum of:
        0.13325372 = weight(_text_:query in 344) [ClassicSimilarity], result of:
          0.13325372 = score(doc=344,freq=4.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.5809541 = fieldWeight in 344, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.0625 = fieldNorm(doc=344)
        0.026746122 = product of:
          0.053492244 = sum of:
            0.053492244 = weight(_text_:22 in 344) [ClassicSimilarity], result of:
              0.053492244 = score(doc=344,freq=2.0), product of:
                0.1728227 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049352113 = queryNorm
                0.30952093 = fieldWeight in 344, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=344)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    In diesem Kapitel wird beschrieben, wie Suchmaschinen Suchanfragen interpretieren können, um letztendlich den Nutzern besser auf ihren Kontext zugeschnittene Ergebnisse liefern zu können. Nach einer Diskussion der Notwendigkeit und der Einsatzmöglichkeiten des Query Understanding wird aufgezeigt, auf welcher Datenbasis und an welchen Ansatzpunkten Suchanfragen interpretiert werden können. Dann erfolgt eine Erläuterung der Interpretationsmöglichkeiten anhand der Suchanfragen-Facetten von Calderon-Benavides et al. (2010), welcher sich eine Diskussion der Verfahren zur Ermittlung der Facetten anschließt.
    Date
    18. 9.2018 18:22:18
  10. Hock, R.E.: How to do field searching in Web search engines : a field trip (1998) 0.10
    0.10378559 = product of:
      0.31135675 = sum of:
        0.31135675 = sum of:
          0.2357073 = weight(_text_:page in 3601) [ClassicSimilarity], result of:
            0.2357073 = score(doc=3601,freq=6.0), product of:
              0.27565226 = queryWeight, product of:
                5.5854197 = idf(docFreq=450, maxDocs=44218)
                0.049352113 = queryNorm
              0.8550893 = fieldWeight in 3601, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                5.5854197 = idf(docFreq=450, maxDocs=44218)
                0.0625 = fieldNorm(doc=3601)
          0.075649455 = weight(_text_:22 in 3601) [ClassicSimilarity], result of:
            0.075649455 = score(doc=3601,freq=4.0), product of:
              0.1728227 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.049352113 = queryNorm
              0.4377287 = fieldWeight in 3601, product of:
                2.0 = tf(freq=4.0), with freq of:
                  4.0 = termFreq=4.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0625 = fieldNorm(doc=3601)
      0.33333334 = coord(1/3)
    
    Abstract
    Explains how 5 Internet search engines (AltaVista, HotBot, InfoSeek, Lycos, and Yahoo) handle field searching. Includes a chart which identifies where on a search engine's page a particular field is searched and the prefix syntax used, and gives examples. Details the individual fields that can be searched: data, title, URL, images, audiovideo and other page content, links and page depth
    Source
    Online. 22(1998) no.3, S.18-22
  11. Kajanan, S.; Bao, Y.; Datta, A.; VanderMeer, D.; Dutta, K.: Efficient automatic search query formulation using phrase-level analysis (2014) 0.10
    0.10210096 = product of:
      0.15315144 = sum of:
        0.09422461 = weight(_text_:query in 1264) [ClassicSimilarity], result of:
          0.09422461 = score(doc=1264,freq=8.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.41079655 = fieldWeight in 1264, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.03125 = fieldNorm(doc=1264)
        0.058926824 = product of:
          0.11785365 = sum of:
            0.11785365 = weight(_text_:page in 1264) [ClassicSimilarity], result of:
              0.11785365 = score(doc=1264,freq=6.0), product of:
                0.27565226 = queryWeight, product of:
                  5.5854197 = idf(docFreq=450, maxDocs=44218)
                  0.049352113 = queryNorm
                0.42754465 = fieldWeight in 1264, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  5.5854197 = idf(docFreq=450, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1264)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Over the past decade, the volume of information available digitally over the Internet has grown enormously. Technical developments in the area of search, such as Google's Page Rank algorithm, have proved so good at serving relevant results that Internet search has become integrated into daily human activity. One can endlessly explore topics of interest simply by querying and reading through the resulting links. Yet, although search engines are well known for providing relevant results based on users' queries, users do not always receive the results they are looking for. Google's Director of Research describes clickstream evidence of frustrated users repeatedly reformulating queries and searching through page after page of results. Given the general quality of search engine results, one must consider the possibility that the frustrated user's query is not effective; that is, it does not describe the essence of the user's interest. Indeed, extensive research into human search behavior has found that humans are not very effective at formulating good search queries that describe what they are interested in. Ideally, the user should simply point to a portion of text that sparked the user's interest, and a system should automatically formulate a search query that captures the essence of the text. In this paper, we describe an implemented system that provides this capability. We first describe how our work differs from existing work in automatic query formulation, and propose a new method for improved quantification of the relevance of candidate search terms drawn from input text using phrase-level analysis. We then propose an implementable method designed to provide relevant queries based on a user's text input. We demonstrate the quality of our results and performance of our system through experimental studies. Our results demonstrate that our system produces relevant search terms with roughly two-thirds precision and recall compared to search terms selected by experts, and that typical users find significantly more relevant results (31% more relevant) more quickly (64% faster) using our system than self-formulated search queries. Further, we show that our implementation can scale to request loads of up to 10 requests per second within current online responsiveness expectations (<2-second response times at the highest loads tested).
  12. Budd, J.M.: ¬The complexity of information retrieval : a hypothetical example (1996) 0.10
    0.10080894 = product of:
      0.15121341 = sum of:
        0.11778076 = weight(_text_:query in 4928) [ClassicSimilarity], result of:
          0.11778076 = score(doc=4928,freq=2.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.5134957 = fieldWeight in 4928, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.078125 = fieldNorm(doc=4928)
        0.03343265 = product of:
          0.0668653 = sum of:
            0.0668653 = weight(_text_:22 in 4928) [ClassicSimilarity], result of:
              0.0668653 = score(doc=4928,freq=2.0), product of:
                0.1728227 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049352113 = queryNorm
                0.38690117 = fieldWeight in 4928, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=4928)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Inquiries made by academic library users are more complex than they may appear. Successful information retrieval based on complex queries is a function of cataloguing, classification, and the librarian's interpretation. Explores aspects of complexitiy using a proposed query as an example
    Source
    Journal of academic librarianship. 22(1996) no.2, S.111-117
  13. Kozikowski, P. et al.: Support of part-whole relations in query answering (2016) 0.10
    0.10080894 = product of:
      0.15121341 = sum of:
        0.11778076 = weight(_text_:query in 2754) [ClassicSimilarity], result of:
          0.11778076 = score(doc=2754,freq=2.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.5134957 = fieldWeight in 2754, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.078125 = fieldNorm(doc=2754)
        0.03343265 = product of:
          0.0668653 = sum of:
            0.0668653 = weight(_text_:22 in 2754) [ClassicSimilarity], result of:
              0.0668653 = score(doc=2754,freq=2.0), product of:
                0.1728227 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049352113 = queryNorm
                0.38690117 = fieldWeight in 2754, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2754)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Date
    1. 2.2016 18:25:22
  14. Park, E.-K.; Ra, D.-Y.; Jang, M.-G.: Techniques for improving web retrieval effectiveness (2005) 0.10
    0.100648284 = product of:
      0.15097243 = sum of:
        0.09994029 = weight(_text_:query in 1060) [ClassicSimilarity], result of:
          0.09994029 = score(doc=1060,freq=4.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.43571556 = fieldWeight in 1060, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.046875 = fieldNorm(doc=1060)
        0.051032126 = product of:
          0.10206425 = sum of:
            0.10206425 = weight(_text_:page in 1060) [ClassicSimilarity], result of:
              0.10206425 = score(doc=1060,freq=2.0), product of:
                0.27565226 = queryWeight, product of:
                  5.5854197 = idf(docFreq=450, maxDocs=44218)
                  0.049352113 = queryNorm
                0.37026453 = fieldWeight in 1060, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.5854197 = idf(docFreq=450, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1060)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This paper talks about several schemes for improving retrieval effectiveness that can be used in the named page finding tasks of web information retrieval (Overview of the TREC-2002 web track. In: Proceedings of the Eleventh Text Retrieval Conference TREC-2002, NIST Special Publication #500-251, 2003). These methods were applied on top of the basic information retrieval model as additional mechanisms to upgrade the system. Use of the title of web pages was found to be effective. It was confirmed that anchor texts of incoming links was beneficial as suggested in other works. Sentence-query similarity is a new type of information proposed by us and was identified to be the best information to take advantage of. Stratifying and re-ranking the retrieval list based on the maximum count of index terms in common between a sentence and a query resulted in significant improvement of performance. To demonstrate these facts a large-scale web information retrieval system was developed and used for experimentation.
  15. Sanchiza, M.; Chinb, J.; Chevaliera, A.; Fuc, W.T.; Amadieua, F.; Hed, J.: Searching for information on the web : impact of cognitive aging, prior domain knowledge and complexity of the search problems (2017) 0.10
    0.100648284 = product of:
      0.15097243 = sum of:
        0.09994029 = weight(_text_:query in 3294) [ClassicSimilarity], result of:
          0.09994029 = score(doc=3294,freq=4.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.43571556 = fieldWeight in 3294, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.046875 = fieldNorm(doc=3294)
        0.051032126 = product of:
          0.10206425 = sum of:
            0.10206425 = weight(_text_:page in 3294) [ClassicSimilarity], result of:
              0.10206425 = score(doc=3294,freq=2.0), product of:
                0.27565226 = queryWeight, product of:
                  5.5854197 = idf(docFreq=450, maxDocs=44218)
                  0.049352113 = queryNorm
                0.37026453 = fieldWeight in 3294, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.5854197 = idf(docFreq=450, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3294)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This study focuses on the impact of age, prior domain knowledge and cognitive abilities on performance, query production and navigation strategies during information searching. Twenty older adults and nineteen young adults had to answer 12 information search problems of varying nature within two domain knowledge: health and manga. In each domain, participants had to perform two simple fact-finding problems (keywords provided and answer directly accessible on the search engine results page), two difficult fact-finding problems (keywords had to be inferred) and two open-ended information search problems (multiple answers possible and navigation necessary). Results showed that prior domain knowledge helped older adults improve navigation (i.e. reduced the number of webpages visited and thus decreased the feeling of disorientation), query production and reformulation (i.e. they formulated semantically more specific queries, and they inferred a greater number of new keywords).
  16. Thenmalar, S.; Geetha, T.V.: Enhanced ontology-based indexing and searching (2014) 0.10
    0.09894903 = product of:
      0.14842354 = sum of:
        0.1367221 = weight(_text_:query in 1633) [ClassicSimilarity], result of:
          0.1367221 = score(doc=1633,freq=22.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.5960754 = fieldWeight in 1633, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1633)
        0.011701428 = product of:
          0.023402857 = sum of:
            0.023402857 = weight(_text_:22 in 1633) [ClassicSimilarity], result of:
              0.023402857 = score(doc=1633,freq=2.0), product of:
                0.1728227 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049352113 = queryNorm
                0.1354154 = fieldWeight in 1633, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1633)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Purpose - The purpose of this paper is to improve the conceptual-based search by incorporating structural ontological information such as concepts and relations. Generally, Semantic-based information retrieval aims to identify relevant information based on the meanings of the query terms or on the context of the terms and the performance of semantic information retrieval is carried out through standard measures-precision and recall. Higher precision leads to the (meaningful) relevant documents obtained and lower recall leads to the less coverage of the concepts. Design/methodology/approach - In this paper, the authors enhance the existing ontology-based indexing proposed by Kohler et al., by incorporating sibling information to the index. The index designed by Kohler et al., contains only super and sub-concepts from the ontology. In addition, in our approach, we focus on two tasks; query expansion and ranking of the expanded queries, to improve the efficiency of the ontology-based search. The aforementioned tasks make use of ontological concepts, and relations existing between those concepts so as to obtain semantically more relevant search results for a given query. Findings - The proposed ontology-based indexing technique is investigated by analysing the coverage of concepts that are being populated in the index. Here, we introduce a new measure called index enhancement measure, to estimate the coverage of ontological concepts being indexed. We have evaluated the ontology-based search for the tourism domain with the tourism documents and tourism-specific ontology. The comparison of search results based on the use of ontology "with and without query expansion" is examined to estimate the efficiency of the proposed query expansion task. The ranking is compared with the ORank system to evaluate the performance of our ontology-based search. From these analyses, the ontology-based search results shows better recall when compared to the other concept-based search systems. The mean average precision of the ontology-based search is found to be 0.79 and the recall is found to be 0.65, the ORank system has the mean average precision of 0.62 and the recall is found to be 0.51, while the concept-based search has the mean average precision of 0.56 and the recall is found to be 0.42. Practical implications - When the concept is not present in the domain-specific ontology, the concept cannot be indexed. When the given query term is not available in the ontology then the term-based results are retrieved. Originality/value - In addition to super and sub-concepts, we incorporate the concepts present in same level (siblings) to the ontological index. The structural information from the ontology is determined for the query expansion. The ranking of the documents depends on the type of the query (single concept query, multiple concept queries and concept with relation queries) and the ontological relations that exists in the query and the documents. With this ontological structural information, the search results showed us better coverage of concepts with respect to the query.
    Date
    20. 1.2015 18:30:22
  17. Friman, J.; Kangaspunta, J.; Leppäniemi, S.; Rasi, P.; Virrankoski, A.: Query performance analyser : a tool for teaching information retrieval skills through an educational game (2005) 0.10
    0.09893282 = product of:
      0.14839922 = sum of:
        0.13168289 = weight(_text_:query in 3010) [ClassicSimilarity], result of:
          0.13168289 = score(doc=3010,freq=10.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.5741056 = fieldWeight in 3010, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3010)
        0.016716326 = product of:
          0.03343265 = sum of:
            0.03343265 = weight(_text_:22 in 3010) [ClassicSimilarity], result of:
              0.03343265 = score(doc=3010,freq=2.0), product of:
                0.1728227 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049352113 = queryNorm
                0.19345059 = fieldWeight in 3010, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3010)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The role of a modern librarian has become more and more demanding in the information age. One of the new challenges for the information specialists is what's usually called "the teaching librarian", meaning that the librarian or information specialist should be able to teach at least basic practical searching skills to the patrons in need for relevant information. Query Performance Analyser (QPA) is a tool for analysing and comparing the performance of individual queries. It has been developed in the department of information studies at the University of Tampere. It can be used in user training to demonstrate the characteristics of IR systems and different searching strategies. Usually users can't get any feedback about the effectiveness of their queries and therefore may have difficulties to perceive the actual fectiveness of a query formulated, or the effect changes between queries. QPA provides a instant visual feedback about the performance of a given query and gives the user a possibility to compare the effectiveness of multiple queries and the performance of different query formulation strategies. QPA is based on predefined search topics. They all contain a corpus of documents that are relevant to the given topic. The purpose of this paper is to give a brief insight to the infrastructure of QPA, the basic :Functionality of the QPA-based game, and to its implementation in IR education.
    Date
    22. 7.2009 11:03:43
  18. Aloteibi, S.; Sanderson, M.: Analyzing geographic query reformulation : an exploratory study (2014) 0.10
    0.09893282 = product of:
      0.14839922 = sum of:
        0.13168289 = weight(_text_:query in 1177) [ClassicSimilarity], result of:
          0.13168289 = score(doc=1177,freq=10.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.5741056 = fieldWeight in 1177, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1177)
        0.016716326 = product of:
          0.03343265 = sum of:
            0.03343265 = weight(_text_:22 in 1177) [ClassicSimilarity], result of:
              0.03343265 = score(doc=1177,freq=2.0), product of:
                0.1728227 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049352113 = queryNorm
                0.19345059 = fieldWeight in 1177, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1177)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Search engine users typically engage in multiquery sessions in their quest to fulfill their information needs. Despite a plethora of research findings suggesting that a significant group of users look for information within a specific geographical scope, existing reformulation studies lack a focused analysis of how users reformulate geographic queries. This study comprehensively investigates the ways in which users reformulate such needs in an attempt to fill this gap in the literature. Reformulated sessions were sampled from a query log of a major search engine to extract 2,400 entries that were manually inspected to filter geo sessions. This filter identified 471 search sessions that included geographical intent, and these sessions were analyzed quantitatively and qualitatively. The results revealed that one in five of the users who reformulated their queries were looking for geographically related information. They reformulated their queries by changing the content of the query rather than the structure. Users were not following a unified sequence of modifications and instead performed a single reformulation action. However, in some cases it was possible to anticipate their next move. A number of tasks in geo modifications were identified, including standard, multi-needs, multi-places, and hybrid approaches. The research concludes that it is important to specialize query reformulation studies to focus on particular query types rather than generically analyzing them, as it is apparent that geographic queries have their special reformulation characteristics.
    Date
    26. 1.2014 18:48:22
  19. Brandão, W.C.; Santos, R.L.T.; Ziviani, N.; Moura, E.S. de; Silva, A.S. da: Learning to expand queries using entities (2014) 0.10
    0.09893282 = product of:
      0.14839922 = sum of:
        0.13168289 = weight(_text_:query in 1343) [ClassicSimilarity], result of:
          0.13168289 = score(doc=1343,freq=10.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.5741056 = fieldWeight in 1343, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1343)
        0.016716326 = product of:
          0.03343265 = sum of:
            0.03343265 = weight(_text_:22 in 1343) [ClassicSimilarity], result of:
              0.03343265 = score(doc=1343,freq=2.0), product of:
                0.1728227 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049352113 = queryNorm
                0.19345059 = fieldWeight in 1343, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1343)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    A substantial fraction of web search queries contain references to entities, such as persons, organizations, and locations. Recently, methods that exploit named entities have been shown to be more effective for query expansion than traditional pseudorelevance feedback methods. In this article, we introduce a supervised learning approach that exploits named entities for query expansion using Wikipedia as a repository of high-quality feedback documents. In contrast with existing entity-oriented pseudorelevance feedback approaches, we tackle query expansion as a learning-to-rank problem. As a result, not only do we select effective expansion terms but we also weigh these terms according to their predicted effectiveness. To this end, we exploit the rich structure of Wikipedia articles to devise discriminative term features, including each candidate term's proximity to the original query terms, as well as its frequency across multiple article fields and in category and infobox descriptors. Experiments on three Text REtrieval Conference web test collections attest the effectiveness of our approach, with gains of up to 23.32% in terms of mean average precision, 19.49% in terms of precision at 10, and 7.86% in terms of normalized discounted cumulative gain compared with a state-of-the-art approach for entity-oriented query expansion.
    Date
    22. 8.2014 17:07:50
  20. Liu, Y.; Zhang, M.; Cen, R.; Ru, L.; Ma, S.: Data cleansing for Web information retrieval using query independent features (2007) 0.10
    0.096351944 = product of:
      0.14452791 = sum of:
        0.10200114 = weight(_text_:query in 607) [ClassicSimilarity], result of:
          0.10200114 = score(doc=607,freq=6.0), product of:
            0.22937049 = queryWeight, product of:
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.049352113 = queryNorm
            0.44470036 = fieldWeight in 607, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.6476326 = idf(docFreq=1151, maxDocs=44218)
              0.0390625 = fieldNorm(doc=607)
        0.04252677 = product of:
          0.08505354 = sum of:
            0.08505354 = weight(_text_:page in 607) [ClassicSimilarity], result of:
              0.08505354 = score(doc=607,freq=2.0), product of:
                0.27565226 = queryWeight, product of:
                  5.5854197 = idf(docFreq=450, maxDocs=44218)
                  0.049352113 = queryNorm
                0.30855376 = fieldWeight in 607, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.5854197 = idf(docFreq=450, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=607)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Understanding what kinds of Web pages are the most useful for Web search engine users is a critical task in Web information retrieval (IR). Most previous works used hyperlink analysis algorithms to solve this problem. However, little research has been focused on query-independent Web data cleansing for Web IR. In this paper, we first provide analysis of the differences between retrieval target pages and ordinary ones based on more than 30 million Web pages obtained from both the Text Retrieval Conference (TREC) and a widely used Chinese search engine, SOGOU (www.sogou.com). We further propose a learning-based data cleansing algorithm for reducing Web pages that are unlikely to be useful for user requests. We found that there exists a large proportion of low-quality Web pages in both the English and the Chinese Web page corpus, and retrieval target pages can be identified using query-independent features and cleansing algorithms. The experimental results showed that our algorithm is effective in reducing a large portion of Web pages with a small loss in retrieval target pages. It makes it possible for Web IR tools to meet a large fraction of users' needs with only a small part of pages on the Web. These results may help Web search engines make better use of their limited storage and computation resources to improve search performance.

Languages

Types

  • a 4271
  • m 445
  • el 298
  • s 180
  • x 47
  • b 41
  • r 28
  • i 25
  • ? 8
  • p 6
  • d 4
  • n 4
  • u 2
  • z 2
  • au 1
  • h 1
  • More… Less…

Themes

Subjects

Classifications