Search (42 results, page 1 of 3)

  • × theme_ss:"Retrievalalgorithmen"
  1. Shiri, A.A.; Revie, C.: Query expansion behavior within a thesaurus-enhanced search environment : a user-centered evaluation (2006) 0.09
    0.086742915 = product of:
      0.21685728 = sum of:
        0.1756729 = weight(_text_:thesaurus in 56) [ClassicSimilarity], result of:
          0.1756729 = score(doc=56,freq=12.0), product of:
            0.28093818 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.060794935 = queryNorm
            0.62530804 = fieldWeight in 56, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.0390625 = fieldNorm(doc=56)
        0.041184373 = weight(_text_:22 in 56) [ClassicSimilarity], result of:
          0.041184373 = score(doc=56,freq=2.0), product of:
            0.2128935 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.060794935 = queryNorm
            0.19345059 = fieldWeight in 56, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0390625 = fieldNorm(doc=56)
      0.4 = coord(2/5)
    
    Abstract
    The study reported here investigated the query expansion behavior of end-users interacting with a thesaurus-enhanced search system on the Web. Two groups, namely academic staff and postgraduate students, were recruited into this study. Data were collected from 90 searches performed by 30 users using the OVID interface to the CAB abstracts database. Data-gathering techniques included questionnaires, screen capturing software, and interviews. The results presented here relate to issues of search-topic and search-term characteristics, number and types of expanded queries, usefulness of thesaurus terms, and behavioral differences between academic staff and postgraduate students in their interaction. The key conclusions drawn were that (a) academic staff chose more narrow and synonymous terms than did postgraduate students, who generally selected broader and related terms; (b) topic complexity affected users' interaction with the thesaurus in that complex topics required more query expansion and search term selection; (c) users' prior topic-search experience appeared to have a significant effect on their selection and evaluation of thesaurus terms; (d) in 50% of the searches where additional terms were suggested from the thesaurus, users stated that they had not been aware of the terms at the beginning of the search; this observation was particularly noticeable in the case of postgraduate students.
    Date
    22. 7.2006 16:32:43
  2. Klas, C.-P.; Fuhr, N.; Schaefer, A.: Evaluating strategic support for information access in the DAFFODIL system (2004) 0.07
    0.06530942 = product of:
      0.16327354 = sum of:
        0.1138523 = weight(_text_:objects in 2419) [ClassicSimilarity], result of:
          0.1138523 = score(doc=2419,freq=2.0), product of:
            0.3231294 = queryWeight, product of:
              5.315071 = idf(docFreq=590, maxDocs=44218)
              0.060794935 = queryNorm
            0.35234275 = fieldWeight in 2419, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.315071 = idf(docFreq=590, maxDocs=44218)
              0.046875 = fieldNorm(doc=2419)
        0.049421243 = weight(_text_:22 in 2419) [ClassicSimilarity], result of:
          0.049421243 = score(doc=2419,freq=2.0), product of:
            0.2128935 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.060794935 = queryNorm
            0.23214069 = fieldWeight in 2419, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.046875 = fieldNorm(doc=2419)
      0.4 = coord(2/5)
    
    Abstract
    The digital library system Daffodil is targeted at strategic support of users during the information search process. For searching, exploring and managing digital library objects it provides user-customisable information seeking patterns over a federation of heterogeneous digital libraries. In this paper evaluation results with respect to retrieval effectiveness, efficiency and user satisfaction are presented. The analysis focuses on strategic support for the scientific work-flow. Daffodil supports the whole work-flow, from data source selection over information seeking to the representation, organisation and reuse of information. By embedding high level search functionality into the scientific work-flow, the user experiences better strategic system support due to a more systematic work process. These ideas have been implemented in Daffodil followed by a qualitative evaluation. The evaluation has been conducted with 28 participants, ranging from information seeking novices to experts. The results are promising, as they support the chosen model.
    Date
    16.11.2008 16:22:48
  3. Rada, R.; Bicknell, E.: Ranking documents with a thesaurus (1989) 0.05
    0.045899622 = product of:
      0.2294981 = sum of:
        0.2294981 = weight(_text_:thesaurus in 6908) [ClassicSimilarity], result of:
          0.2294981 = score(doc=6908,freq=2.0), product of:
            0.28093818 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.060794935 = queryNorm
            0.81689894 = fieldWeight in 6908, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.125 = fieldNorm(doc=6908)
      0.2 = coord(1/5)
    
  4. Rada, R.; Barlow, J.; Potharst, J.; Zanstra, P.; Bijstra, D.: Document ranking using an enriched thesaurus (1991) 0.04
    0.042161494 = product of:
      0.21080747 = sum of:
        0.21080747 = weight(_text_:thesaurus in 6626) [ClassicSimilarity], result of:
          0.21080747 = score(doc=6626,freq=12.0), product of:
            0.28093818 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.060794935 = queryNorm
            0.7503696 = fieldWeight in 6626, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.046875 = fieldNorm(doc=6626)
      0.2 = coord(1/5)
    
    Abstract
    A thesaurus may be viewed as a graph, and document retrieval algorithms can exploit this graph when both the documents and the query are represented by thesaurus terms. These retrieval algorithms measure the distance between the query and documents by using the path lengths in the graph. Previous work witj such strategies has shown that the hierarchical relations in the thesaurus are useful but the non-hierarchical are not. This paper shows that when the query explicitly mentions a particular non-hierarchical relation, the retrieval algorithm benefits from the presence of such relations in the thesaurus. Our algorithms were applied to the Excerpta Medica bibliographic citation database whose citations are indexed with terms from the EMTREE thesaurus. We also created an enriched EMTREE by systematically adding non-hierarchical relations from a medical knowledge base. Our algorithms used at one time EMTREE and, at another time, the enriched EMTREE in the course of ranking documents from Excerpta Medica against queries. When, and only when, the query specifically mentioned a particular non-hierarchical relation type, did EMTREE enriched with that relation type lead to a ranking that better corresponded to an expert's ranking
  5. Kulyukin, V.A.; Settle, A.: Ranked retrieval with semantic networks and vector spaces (2001) 0.03
    0.030360613 = product of:
      0.15180306 = sum of:
        0.15180306 = weight(_text_:objects in 6934) [ClassicSimilarity], result of:
          0.15180306 = score(doc=6934,freq=2.0), product of:
            0.3231294 = queryWeight, product of:
              5.315071 = idf(docFreq=590, maxDocs=44218)
              0.060794935 = queryNorm
            0.46979034 = fieldWeight in 6934, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.315071 = idf(docFreq=590, maxDocs=44218)
              0.0625 = fieldNorm(doc=6934)
      0.2 = coord(1/5)
    
    Abstract
    The equivalence of semantic networks with spreading activation and vector spaces with dot product is investigated under ranked retrieval. Semantic networks are viewed as networks of concepts organized in terms of abstraction and packaging relations. It is shown that the two models can be effectively constructed from each other. A formal method is suggested to analyze the models in terms of their relative performance in the same universe of objects
  6. Srinivasan, P.: Intelligent information retrieval using rough set approximations (1989) 0.03
    0.026565535 = product of:
      0.13282767 = sum of:
        0.13282767 = weight(_text_:objects in 2526) [ClassicSimilarity], result of:
          0.13282767 = score(doc=2526,freq=2.0), product of:
            0.3231294 = queryWeight, product of:
              5.315071 = idf(docFreq=590, maxDocs=44218)
              0.060794935 = queryNorm
            0.41106653 = fieldWeight in 2526, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.315071 = idf(docFreq=590, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2526)
      0.2 = coord(1/5)
    
    Abstract
    The theory of rough sets was introduced in 1982. It allows the classification of objects into sets of equivalent members based on their attributes. Any combination of the same objetcts (or even their attributes) may be examined using the resultant classification. The theory has direct applications in the design and evaluation of classification schemes and the selection of discriminating attributes. Introductory papers discuss its application in the domain of medical diagnostic systems and the design of information retrieval systems accessing collections of documents. Advantages offered by the theory are: the implicit inclusion of Boolean logic; term weighting; and the ability to rank retrieved documents.
  7. Voorhees, E.M.: Implementing agglomerative hierarchic clustering algorithms for use in document retrieval (1986) 0.03
    0.026357997 = product of:
      0.13178998 = sum of:
        0.13178998 = weight(_text_:22 in 402) [ClassicSimilarity], result of:
          0.13178998 = score(doc=402,freq=2.0), product of:
            0.2128935 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.060794935 = queryNorm
            0.61904186 = fieldWeight in 402, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.125 = fieldNorm(doc=402)
      0.2 = coord(1/5)
    
    Source
    Information processing and management. 22(1986) no.6, S.465-476
  8. Chen, H.; Lally, A.M.; Zhu, B.; Chau, M.: HelpfulMed : Intelligent searching for medical information over the Internet (2003) 0.02
    0.0248439 = product of:
      0.12421949 = sum of:
        0.12421949 = weight(_text_:thesaurus in 1615) [ClassicSimilarity], result of:
          0.12421949 = score(doc=1615,freq=6.0), product of:
            0.28093818 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.060794935 = queryNorm
            0.44215953 = fieldWeight in 1615, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1615)
      0.2 = coord(1/5)
    
    Abstract
    The Medical professionals and researchers need information from reputable sources to accomplish their work. Unfortunately, the Web has a large number of documents that are irrelevant to their work, even those documents that purport to be "medically-related." This paper describes an architecture designed to integrate advanced searching and indexing algorithms, an automatic thesaurus, or "concept space," and Kohonen-based Self-Organizing Map (SOM) technologies to provide searchers with finegrained results. Initial results indicate that these systems provide complementary retrieval functionalities. HelpfulMed not only allows users to search Web pages and other online databases, but also allows them to build searches through the use of an automatic thesaurus and browse a graphical display of medical-related topics. Evaluation results for each of the different components are included. Our spidering algorithm outperformed both breadth-first search and PageRank spiders an a test collection of 100,000 Web pages. The automatically generated thesaurus performed as well as both MeSH and UMLS-systems which require human mediation for currency. Lastly, a variant of the Kohonen SOM was comparable to MeSH terms in perceived cluster precision and significantly better at perceived cluster recall.
  9. Gauch, S.; Smith, J.B.: ¬An expert system for automatic query reformation (1993) 0.02
    0.024341952 = product of:
      0.12170976 = sum of:
        0.12170976 = weight(_text_:thesaurus in 3693) [ClassicSimilarity], result of:
          0.12170976 = score(doc=3693,freq=4.0), product of:
            0.28093818 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.060794935 = queryNorm
            0.4332261 = fieldWeight in 3693, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.046875 = fieldNorm(doc=3693)
      0.2 = coord(1/5)
    
    Abstract
    Unfamiliarity with search tactics creates difficulties for many users of online retrieval systems. User observations indicate that even experienced searchers use vocabulary incorrectly and rarely reformulate their queries. To address these problems, an expert system for online search assistance was developed. This prototype automatically reformulates queries to improve the search results, and ranks the retrieved passages to speed the identification of relevant information. User's search performance using the expert system was compared with their search performance using an online thesaurus. The following conclusions were reached: (1) the expert system significantly reduced the number of queries necessary to find relevant passages compared with the user searching alone or with the thesaurus. (2) The expert system produced marginally significant improvements in precision compared with the user searching on their own. There was no significant difference in the recall achieved by the three system configurations. (3) Overall, the expert system ranked relevant passages above irrelevant passages
  10. Smeaton, A.F.; Rijsbergen, C.J. van: ¬The retrieval effects of query expansion on a feedback document retrieval system (1983) 0.02
    0.023063248 = product of:
      0.11531624 = sum of:
        0.11531624 = weight(_text_:22 in 2134) [ClassicSimilarity], result of:
          0.11531624 = score(doc=2134,freq=2.0), product of:
            0.2128935 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.060794935 = queryNorm
            0.5416616 = fieldWeight in 2134, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.109375 = fieldNorm(doc=2134)
      0.2 = coord(1/5)
    
    Date
    30. 3.2001 13:32:22
  11. Back, J.: ¬An evaluation of relevancy ranking techniques used by Internet search engines (2000) 0.02
    0.023063248 = product of:
      0.11531624 = sum of:
        0.11531624 = weight(_text_:22 in 3445) [ClassicSimilarity], result of:
          0.11531624 = score(doc=3445,freq=2.0), product of:
            0.2128935 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.060794935 = queryNorm
            0.5416616 = fieldWeight in 3445, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.109375 = fieldNorm(doc=3445)
      0.2 = coord(1/5)
    
    Date
    25. 8.2005 17:42:22
  12. Nakkouzi, Z.S.; Eastman, C.M.: Query formulation for handling negation in information retrieval systems (1990) 0.02
    0.022949811 = product of:
      0.11474905 = sum of:
        0.11474905 = weight(_text_:thesaurus in 3531) [ClassicSimilarity], result of:
          0.11474905 = score(doc=3531,freq=2.0), product of:
            0.28093818 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.060794935 = queryNorm
            0.40844947 = fieldWeight in 3531, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.0625 = fieldNorm(doc=3531)
      0.2 = coord(1/5)
    
    Abstract
    Queries containing negation are widely recognised as presenting problems for both users and systems. In information retrieval systems such problems usually manifest themselves in the use of the NOT operator. Describes an algorithm to transform Boolean queries with negated terms into queries without negation; the transformation process is based on the use of a hierarchical thesaurus. Examines a set of user requests submitted to the Thomas Cooper Library at the University of South Carolina to determine the pattern and frequency of use of negation.
  13. Fuhr, N.: Ranking-Experimente mit gewichteter Indexierung (1986) 0.02
    0.019768497 = product of:
      0.09884249 = sum of:
        0.09884249 = weight(_text_:22 in 58) [ClassicSimilarity], result of:
          0.09884249 = score(doc=58,freq=2.0), product of:
            0.2128935 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.060794935 = queryNorm
            0.46428138 = fieldWeight in 58, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.09375 = fieldNorm(doc=58)
      0.2 = coord(1/5)
    
    Date
    14. 6.2015 22:12:44
  14. Fuhr, N.: Rankingexperimente mit gewichteter Indexierung (1986) 0.02
    0.019768497 = product of:
      0.09884249 = sum of:
        0.09884249 = weight(_text_:22 in 2051) [ClassicSimilarity], result of:
          0.09884249 = score(doc=2051,freq=2.0), product of:
            0.2128935 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.060794935 = queryNorm
            0.46428138 = fieldWeight in 2051, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.09375 = fieldNorm(doc=2051)
      0.2 = coord(1/5)
    
    Date
    14. 6.2015 22:12:56
  15. Calegari, S.; Sanchez, E.: Object-fuzzy concept network : an enrichment of ontologies in semantic information retrieval (2008) 0.02
    0.018975385 = product of:
      0.09487692 = sum of:
        0.09487692 = weight(_text_:objects in 2393) [ClassicSimilarity], result of:
          0.09487692 = score(doc=2393,freq=2.0), product of:
            0.3231294 = queryWeight, product of:
              5.315071 = idf(docFreq=590, maxDocs=44218)
              0.060794935 = queryNorm
            0.29361898 = fieldWeight in 2393, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.315071 = idf(docFreq=590, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2393)
      0.2 = coord(1/5)
    
    Abstract
    This article shows how a fuzzy ontology-based approach can improve semantic documents retrieval. After formally defining a fuzzy ontology and a fuzzy knowledge base, a special type of new fuzzy relationship called (semantic) correlation, which links the concepts or entities in a fuzzy ontology, is discussed. These correlations, first assigned by experts, are updated after querying or when a document has been inserted into a database. Moreover, in order to define a dynamic knowledge of a domain adapting itself to the context, it is shown how to handle a tradeoff between the correct definition of an object, taken in the ontology structure, and the actual meaning assigned by individuals. The notion of a fuzzy concept network is extended, incorporating database objects so that entities and documents can similarly be represented in the network. Information retrieval (IR) algorithm, using an object-fuzzy concept network (O-FCN), is introduced and described. This algorithm allows us to derive a unique path among the entities involved in the query to obtain maxima semantic associations in the knowledge domain. Finally, the study has been validated by querying a database using fuzzy recall, fuzzy precision, and coefficient variant measures in the crisp and fuzzy cases.
  16. Schaefer, A.; Jordan, M.; Klas, C.-P.; Fuhr, N.: Active support for query formulation in virtual digital libraries : a case study with DAFFODIL (2005) 0.02
    0.018975385 = product of:
      0.09487692 = sum of:
        0.09487692 = weight(_text_:objects in 4296) [ClassicSimilarity], result of:
          0.09487692 = score(doc=4296,freq=2.0), product of:
            0.3231294 = queryWeight, product of:
              5.315071 = idf(docFreq=590, maxDocs=44218)
              0.060794935 = queryNorm
            0.29361898 = fieldWeight in 4296, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.315071 = idf(docFreq=590, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4296)
      0.2 = coord(1/5)
    
    Abstract
    Daffodil is a front-end to federated, heterogeneous digital libraries targeting at strategic support of users during the information seeking process. This is done by offering a variety of functions for searching, exploring and managing digital library objects. However, the distributed search increases response time and the conceptual model of the underlying search processes is inherently weaker. This makes query formulation harder and the resulting waiting times can be frustrating. In this paper, we investigate the concept of proactive support during the user's query formulation. For improving user efficiency and satisfaction, we implemented annotations, proactive support and error markers on the query form itself. These functions decrease the probability for syntactical or semantical errors in queries. Furthermore, the user is able to make better tactical decisions and feels more confident that the system handles the query properly. Evaluations with 30 subjects showed that user satisfaction is improved, whereas no conclusive results were received for efficiency.
  17. Ding, Y.; Chowdhury, G.; Foo, S.: Organsising keywords in a Web search environment : a methodology based on co-word analysis (2000) 0.02
    0.01721236 = product of:
      0.08606179 = sum of:
        0.08606179 = weight(_text_:thesaurus in 105) [ClassicSimilarity], result of:
          0.08606179 = score(doc=105,freq=2.0), product of:
            0.28093818 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.060794935 = queryNorm
            0.30633712 = fieldWeight in 105, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.046875 = fieldNorm(doc=105)
      0.2 = coord(1/5)
    
    Abstract
    The rapid development of the Internet and World Wide Web has caused some critical problem for information retrieval. Researchers have made several attempts to solve these problems. Thesauri and subject heading lists as traditional information retrieval tools have been criticised for their efficiency to tackle these newly emerging problems. This paper proposes an information retrieval tool generated by cocitation analysis, comprising keyword clusters with relationships based on the co-occurrences of keywords in the literature. Such a tool can play the role of an associative thesaurus that can provide information about the keywords in a domain that might be useful for information searching and query expansion
  18. Information retrieval : data structures and algorithms (1992) 0.01
    0.014343631 = product of:
      0.07171816 = sum of:
        0.07171816 = weight(_text_:thesaurus in 3495) [ClassicSimilarity], result of:
          0.07171816 = score(doc=3495,freq=2.0), product of:
            0.28093818 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.060794935 = queryNorm
            0.2552809 = fieldWeight in 3495, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3495)
      0.2 = coord(1/5)
    
    Content
    An edited volume containing data structures and algorithms for information retrieval including a disk with examples written in C. for prgrammers and students interested in parsing text, automated indexing, its the first collection in book form of the basic data structures and algorithms that are critical to the storage and retrieval of documents. ------------------Enthält die Kapitel: FRAKES, W.B.: Introduction to information storage and retrieval systems; BAEZA-YATES, R.S.: Introduction to data structures and algorithms related to information retrieval; HARMAN, D. u.a.: Inverted files; FALOUTSOS, C.: Signature files; GONNET, G.H. u.a.: New indices for text: PAT trees and PAT arrays; FORD, D.A. u. S. CHRISTODOULAKIS: File organizations for optical disks; FOX, C.: Lexical analysis and stoplists; FRAKES, W.B.: Stemming algorithms; SRINIVASAN, P.: Thesaurus construction; BAEZA-YATES, R.A.: String searching algorithms; HARMAN, D.: Relevance feedback and other query modification techniques; WARTIK, S.: Boolean operators; WARTIK, S. u.a.: Hashing algorithms; HARMAN, D.: Ranking algorithms; FOX, E.: u.a.: Extended Boolean models; RASMUSSEN, E.: Clustering algorithms; HOLLAAR, L.: Special-purpose hardware for information retrieval; STANFILL, C.: Parallel information retrieval algorithms
  19. Watters, C.; Amoudi, A.: Geosearcher : location-based ranking of search engine results (2003) 0.01
    0.014343631 = product of:
      0.07171816 = sum of:
        0.07171816 = weight(_text_:thesaurus in 5152) [ClassicSimilarity], result of:
          0.07171816 = score(doc=5152,freq=2.0), product of:
            0.28093818 = queryWeight, product of:
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.060794935 = queryNorm
            0.2552809 = fieldWeight in 5152, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.6210785 = idf(docFreq=1182, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5152)
      0.2 = coord(1/5)
    
    Abstract
    Waters and Amoudi describe GeoSearcher, a prototype ranking program that arranges search engine results along a geo-spatial dimension without the provision of geo-spatial meta-tags or the use of geo-spatial feature extraction. GeoSearcher uses URL analysis, IptoLL, Whois, and the Getty Thesaurus of Geographic Names to determine site location. It accepts the first 200 sites returned by a search engine, identifies the coordinates, calculates their distance from a reference point and ranks in ascending order by this value. For any retrieved site the system checks if it has already been located in the current session, then sends the domain name to Whois to generate a return of a two letter country code and an area code. With no success the name is stripped one level and resent. If this fails the top level domain is tested for being a country code. Any remaining unmatched names go to IptoLL. Distance is calculated using the center point of the geographic area and a provided reference location. A test run on a set of 100 URLs from a search was successful in locating 90 sites. Eighty three pages could be manually found and 68 had sufficient information to verify location determination. Of these 65 ( 95%) had been assigned reasonably correct geographic locations. A random set of URLs used instead of a search result, yielded 80% success.
  20. MacFarlane, A.; Robertson, S.E.; McCann, J.A.: Parallel computing for passage retrieval (2004) 0.01
    0.013178999 = product of:
      0.06589499 = sum of:
        0.06589499 = weight(_text_:22 in 5108) [ClassicSimilarity], result of:
          0.06589499 = score(doc=5108,freq=2.0), product of:
            0.2128935 = queryWeight, product of:
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.060794935 = queryNorm
            0.30952093 = fieldWeight in 5108, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5018296 = idf(docFreq=3622, maxDocs=44218)
              0.0625 = fieldNorm(doc=5108)
      0.2 = coord(1/5)
    
    Date
    20. 1.2007 18:30:22

Years

Languages

  • e 38
  • d 4

Types

  • a 38
  • m 2
  • el 1
  • r 1
  • s 1
  • More… Less…