Search (105 results, page 1 of 6)

  • × theme_ss:"Semantisches Umfeld in Indexierung u. Retrieval"
  1. Bhogal, J.; Macfarlane, A.; Smith, P.: ¬A review of ontology based query expansion (2007) 0.05
    0.052956074 = product of:
      0.12356417 = sum of:
        0.036320645 = weight(_text_:management in 919) [ClassicSimilarity], result of:
          0.036320645 = score(doc=919,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.2606825 = fieldWeight in 919, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.0546875 = fieldNorm(doc=919)
        0.061792135 = weight(_text_:case in 919) [ClassicSimilarity], result of:
          0.061792135 = score(doc=919,freq=2.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.34001783 = fieldWeight in 919, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.0546875 = fieldNorm(doc=919)
        0.02545139 = product of:
          0.05090278 = sum of:
            0.05090278 = weight(_text_:studies in 919) [ClassicSimilarity], result of:
              0.05090278 = score(doc=919,freq=2.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.30860704 = fieldWeight in 919, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=919)
          0.5 = coord(1/2)
      0.42857143 = coord(3/7)
    
    Abstract
    This paper examines the meaning of context in relation to ontology based query expansion and contains a review of query expansion approaches. The various query expansion approaches include relevance feedback, corpus dependent knowledge models and corpus independent knowledge models. Case studies detailing query expansion using domain-specific and domain-independent ontologies are also included. The penultimate section attempts to synthesise the information obtained from the review and provide success factors in using an ontology for query expansion. Finally the area of further research in applying context from an ontology to query expansion within a newswire domain is described.
    Source
    Information processing and management. 43(2007) no.4, S.866-886
  2. Case, D.O.: Looking for information : a survey on research on information seeking, needs, and behavior (2002) 0.03
    0.03173722 = product of:
      0.11108026 = sum of:
        0.07490338 = weight(_text_:case in 1270) [ClassicSimilarity], result of:
          0.07490338 = score(doc=1270,freq=16.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.41216385 = fieldWeight in 1270, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1270)
        0.03617688 = product of:
          0.07235376 = sum of:
            0.07235376 = weight(_text_:studies in 1270) [ClassicSimilarity], result of:
              0.07235376 = score(doc=1270,freq=22.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.43865734 = fieldWeight in 1270, product of:
                  4.690416 = tf(freq=22.0), with freq of:
                    22.0 = termFreq=22.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=1270)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Footnote
    Rez. in: JASIST 54(2003) no.7, S.695-697 (R. Savolainen): "Donald O. Case has written an ambitious book to create an overall picture of the major approaches to information needs and seeking (INS) studies. The aim to write an extensive review is reflected in the list of references containing about 700 items. The high ambitions are explained an p. 14, where Case states that he is aiming at a multidisciplinary understanding of the concept of information seeking. In the Preface, the author characterizes his book as an introduction to the topic for students at the graduate level, as well as as a review and handbook for scholars engagged in information behavior research. In my view, Looking for Information is particularly welcome as an academic textbook because the field of INS studies suffers from the lack of monographs. Along with the continuous growth of the number of journal articles and conference papers, there is a genuine need for a book that picks up the numerous pieces and puts them together. The use of the study as a textbook is facilitated by clearly delineated sections an major themes and the wealth of concrete examples of information seeking in everyday contexts. The book is lucidly written and it is accessible to novice readers, too. At first glance, the idea of providing a comprehensive review of INS studies may seem a mission impossible because the current number of articles, papers, and other contributions in this field is nearing the 10,000 range (p. 224). Donald Case is not alone in the task of coming to grips with an increasing number of studies; similar problems have been faced by those writing INS-related chapters for the Annual Review of Information Science and Technology (ARIST). Case has solved the problem of "too many publications to be reviewed" by concentrating an the INS literature published during the last two decades. Secondly, studies an library use and information retrieval are discussed only to a limited extent. In addition, Case is highly selective as to studies focusing an the use of specific sources and channels such as WWW. These delineations are reasonable, even though they beg some questions. First, how should one draw the line between studies an information seeking and information retrieval? Case does not discuss this question in greater detail, although in recent years, the overlapping areas of information seeking and retrieval studies have been broadened, along with the growing importance of WWW in information seeking/retrieval. Secondly, how can one define the concept of information searching (or, more specifically, Internet or Web searching) in relation to information seeking and information retrieval? In the field of Web searching studies, there is an increasing number of contributions that are of direct relevance to information-seeking studies. Clearly, the advent of the Internet, particularly, the Web, has blurred the previous lines between INS and IR literature, making them less clear cut. The book consists of five main sections, and comprises 13 chapters. There is an Appendix serving the needs of an INS textbook (questions for discussion and application). The structure of the book is meticulously planned and, as a whole, it offers a sufficiently balanced contribution to theoretical, methodological, and empirical issues of INS. The title, Looking for Information: A Survey of Research an Information Seeking, Needs, and Behavior aptly describes the main substance of the book. . . . It is easy to agree with Case about the significance of the problem of specialization and fragmentation. This problem seems to be concomitant with the broadening field of INS research. In itself, Case's book can be interpreted as a struggle against this fragmentation. His book suggests that this struggle is not hopeless and that it is still possible to draw an overall picture of the evolving research field. The major pieces of the puzzle were found and the book will provide a useful overview of INS studies for many years."
  3. Shiri, A.A.; Revie, C.; Chowdhury, G.: Thesaurus-assisted search term selection and query expansion : a review of user-centred studies (2002) 0.03
    0.030400421 = product of:
      0.10640147 = sum of:
        0.052964687 = weight(_text_:case in 1330) [ClassicSimilarity], result of:
          0.052964687 = score(doc=1330,freq=2.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.29144385 = fieldWeight in 1330, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.046875 = fieldNorm(doc=1330)
        0.053436786 = product of:
          0.10687357 = sum of:
            0.10687357 = weight(_text_:studies in 1330) [ClassicSimilarity], result of:
              0.10687357 = score(doc=1330,freq=12.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.6479398 = fieldWeight in 1330, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1330)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    This paper provides a review of the literature related to the application of domain-specific thesauri in the search and retrieval process. Focusing an studies that adopt a user-centred approach, the review presents a survey of the methodologies and results from empirical studies undertaken an the use of thesauri as sources of term selection for query formulation and expansion during the search process. It summarises the ways in which domain-specific thesauri from different disciplines have been used by various types of users and how these tools aid users in the selection of search terms. The review consists of two main sections: first, studies an thesaurus-aided search term selection; and second, studies dealing with query expansion using thesauri. Both sections are illustrated with case studies that have adopted a user-centred approach.
  4. Lin, J.; DiCuccio, M.; Grigoryan, V.; Wilbur, W.J.: Navigating information spaces : a case study of related article search in PubMed (2008) 0.03
    0.030295819 = product of:
      0.10603536 = sum of:
        0.031131983 = weight(_text_:management in 2124) [ClassicSimilarity], result of:
          0.031131983 = score(doc=2124,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.22344214 = fieldWeight in 2124, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.046875 = fieldNorm(doc=2124)
        0.07490338 = weight(_text_:case in 2124) [ClassicSimilarity], result of:
          0.07490338 = score(doc=2124,freq=4.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.41216385 = fieldWeight in 2124, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.046875 = fieldNorm(doc=2124)
      0.2857143 = coord(2/7)
    
    Abstract
    The concept of an "information space" provides a powerful metaphor for guiding the design of interactive retrieval systems. We present a case study of related article search, a browsing tool designed to help users navigate the information space defined by results of the PubMed® search engine. This feature leverages content-similarity links that tie MEDLINE® citations together in a vast document network. We examine the effectiveness of related article search from two perspectives: a topological analysis of networks generated from information needs represented in the TREC 2005 genomics track and a query log analysis of real PubMed users. Together, data suggest that related article search is a useful feature and that browsing related articles has become an integral part of how users interact with PubMed.
    Source
    Information processing and management. 44(2008) no.5, S.1771-1783
  5. Vidinli, I.B.; Ozcan, R.: New query suggestion framework and algorithms : a case study for an educational search engine (2016) 0.03
    0.030295819 = product of:
      0.10603536 = sum of:
        0.031131983 = weight(_text_:management in 3185) [ClassicSimilarity], result of:
          0.031131983 = score(doc=3185,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.22344214 = fieldWeight in 3185, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.046875 = fieldNorm(doc=3185)
        0.07490338 = weight(_text_:case in 3185) [ClassicSimilarity], result of:
          0.07490338 = score(doc=3185,freq=4.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.41216385 = fieldWeight in 3185, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.046875 = fieldNorm(doc=3185)
      0.2857143 = coord(2/7)
    
    Abstract
    Query suggestion is generally an integrated part of web search engines. In this study, we first redefine and reduce the query suggestion problem as "comparison of queries". We then propose a general modular framework for query suggestion algorithm development. We also develop new query suggestion algorithms which are used in our proposed framework, exploiting query, session and user features. As a case study, we use query logs of a real educational search engine that targets K-12 students in Turkey. We also exploit educational features (course, grade) in our query suggestion algorithms. We test our framework and algorithms over a set of queries by an experiment and demonstrate a 66-90% statistically significant increase in relevance of query suggestions compared to a baseline method.
    Source
    Information processing and management. 52(2016) no.5, S.733-752
  6. Nie, J.-Y.: Query expansion and query translation as logical inference (2003) 0.02
    0.023947554 = product of:
      0.08381643 = sum of:
        0.052964687 = weight(_text_:case in 1425) [ClassicSimilarity], result of:
          0.052964687 = score(doc=1425,freq=2.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.29144385 = fieldWeight in 1425, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.046875 = fieldNorm(doc=1425)
        0.030851744 = product of:
          0.06170349 = sum of:
            0.06170349 = weight(_text_:studies in 1425) [ClassicSimilarity], result of:
              0.06170349 = score(doc=1425,freq=4.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.37408823 = fieldWeight in 1425, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1425)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    A number of studies have examined the problems of query expansion in monolingual Information Retrieval (IR), and query translation for crosslanguage IR. However, no link has been made between them. This article first shows that query translation is a special case of query expansion. There is also another set of studies an inferential IR. Again, there is no relationship established with query translation or query expansion. The second claim of this article is that logical inference is a general form that covers query expansion and query translation. This analysis provides a unified view of different subareas of IR. We further develop the inferential IR approach in two particular contexts: using fuzzy logic and probability theory. The evaluation formulas obtained are shown to strongly correspond to those used in other IR models. This indicates that inference is indeed the core of advanced IR.
  7. Salaba, A.; Zeng, M.L.: Extending the "Explore" user task beyond subject authority data into the linked data sphere (2014) 0.02
    0.023255412 = product of:
      0.081393935 = sum of:
        0.061792135 = weight(_text_:case in 1465) [ClassicSimilarity], result of:
          0.061792135 = score(doc=1465,freq=2.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.34001783 = fieldWeight in 1465, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1465)
        0.019601801 = product of:
          0.039203603 = sum of:
            0.039203603 = weight(_text_:22 in 1465) [ClassicSimilarity], result of:
              0.039203603 = score(doc=1465,freq=2.0), product of:
                0.14475311 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.041336425 = queryNorm
                0.2708308 = fieldWeight in 1465, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1465)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    "Explore" is a user task introduced in the Functional Requirements for Subject Authority Data (FRSAD) final report. Through various case scenarios, the authors discuss how structured data, presented based on Linked Data principles and using knowledge organisation systems (KOS) as the backbone, extend the explore task within and beyond subject authority data.
    Source
    Knowledge organization in the 21st century: between historical patterns and future prospects. Proceedings of the Thirteenth International ISKO Conference 19-22 May 2014, Kraków, Poland. Ed.: Wieslaw Babik
  8. Bayer, O.; Höhfeld, S.; Josbächer, F.; Kimm, N.; Kradepohl, I.; Kwiatkowski, M.; Puschmann, C.; Sabbagh, M.; Werner, N.; Vollmer, U.: Evaluation of an ontology-based knowledge-management-system : a case study of Convera RetrievalWare 8.0 (2005) 0.02
    0.023093324 = product of:
      0.08082663 = sum of:
        0.036689393 = weight(_text_:management in 624) [ClassicSimilarity], result of:
          0.036689393 = score(doc=624,freq=4.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.2633291 = fieldWeight in 624, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.0390625 = fieldNorm(doc=624)
        0.04413724 = weight(_text_:case in 624) [ClassicSimilarity], result of:
          0.04413724 = score(doc=624,freq=2.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.24286987 = fieldWeight in 624, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.0390625 = fieldNorm(doc=624)
      0.2857143 = coord(2/7)
    
    Abstract
    With RetrievalWare 8.0(TM) the American company Convera offers an elaborated software in the range of Information Retrieval, Information Indexing and Knowledge Management. Convera promises the possibility of handling different file formats in many different languages. Regarding comparable products one innovation is to be stressed particularly: the possibility of the preparation as well as integration of an ontology. One tool of the software package is useful in order to produce ontologies manually, to process existing ontologies and to import the very. The processing of search results is also to be mentioned. By means of categorization strategies search results can be classified dynamically and presented in personalized representations. This study presents an evaluation of the functions and components of the system. Technological aspects and modes of operation under the surface of Convera RetrievalWare will be analysed, with a focus on the creation of libraries and thesauri, and the problems posed by the integration of an existing thesaurus. Broader aspects such as usability and system ergonomics are integrated in the examination as well.
  9. Looking for information : a survey on research on information seeking, needs, and behavior (2016) 0.02
    0.022998964 = product of:
      0.08049637 = sum of:
        0.04413724 = weight(_text_:case in 3803) [ClassicSimilarity], result of:
          0.04413724 = score(doc=3803,freq=2.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.24286987 = fieldWeight in 3803, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3803)
        0.03635913 = product of:
          0.07271826 = sum of:
            0.07271826 = weight(_text_:studies in 3803) [ClassicSimilarity], result of:
              0.07271826 = score(doc=3803,freq=8.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.44086722 = fieldWeight in 3803, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3803)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    The 4th edition of this popular and well-cited text is now co-authored, and includes significant changes from earlier texts. Presenting a comprehensive review of over a century of research on information behavior (IB), this book is intended for students in information studies and disciplines interested in research on information activities. The initial two chapters introduce IB as a multi-disciplinary topic, the 3rd provides a brief history of research on information seeking. Chapter four discusses what is meant by the terms "information" and "knowledge. "Chapter five discusses "information needs," and how they are addressed. The 6th chapter identifies many related concepts. Twelve models of information behavior (expanded from earlier editions) are illustrated in chapter seven. Chapter eight reviews various paradigms and theories informing IB research. Chapter nine examines research methods invoked in IB studies and a discussion of qualitative and mixed approaches. The 10th chapter gives examples of IB studies by context. The final chapter looks at strengths and weaknesses, recent trends, and future development.
    Editor
    Case, D.O. u. Lisa M. Given
    Series
    Studies in information
  10. Wongthontham, P.; Abu-Salih, B.: Ontology-based approach for semantic data extraction from social big data : state-of-the-art and research directions (2018) 0.02
    0.021365764 = product of:
      0.074780166 = sum of:
        0.052964687 = weight(_text_:case in 4097) [ClassicSimilarity], result of:
          0.052964687 = score(doc=4097,freq=2.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.29144385 = fieldWeight in 4097, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.046875 = fieldNorm(doc=4097)
        0.021815477 = product of:
          0.043630954 = sum of:
            0.043630954 = weight(_text_:studies in 4097) [ClassicSimilarity], result of:
              0.043630954 = score(doc=4097,freq=2.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.26452032 = fieldWeight in 4097, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4097)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    A challenge of managing and extracting useful knowledge from social media data sources has attracted much attention from academic and industry. To address this challenge, semantic analysis of textual data is focused in this paper. We propose an ontology-based approach to extract semantics of textual data and define the domain of data. In other words, we semantically analyse the social data at two levels i.e. the entity level and the domain level. We have chosen Twitter as a social channel challenge for a purpose of concept proof. Domain knowledge is captured in ontologies which are then used to enrich the semantics of tweets provided with specific semantic conceptual representation of entities that appear in the tweets. Case studies are used to demonstrate this approach. We experiment and evaluate our proposed approach with a public dataset collected from Twitter and from the politics domain. The ontology-based approach leverages entity extraction and concept mappings in terms of quantity and accuracy of concept identification.
  11. Sacco, G.M.: Dynamic taxonomies and guided searches (2006) 0.02
    0.018297652 = product of:
      0.06404178 = sum of:
        0.036320645 = weight(_text_:management in 5295) [ClassicSimilarity], result of:
          0.036320645 = score(doc=5295,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.2606825 = fieldWeight in 5295, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.0546875 = fieldNorm(doc=5295)
        0.027721135 = product of:
          0.05544227 = sum of:
            0.05544227 = weight(_text_:22 in 5295) [ClassicSimilarity], result of:
              0.05544227 = score(doc=5295,freq=4.0), product of:
                0.14475311 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.041336425 = queryNorm
                0.38301262 = fieldWeight in 5295, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5295)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    A new search paradigm, in which the primary user activity is the guided exploration of a complex information space rather than the retrieval of items based on precise specifications, is proposed. The author claims that this paradigm is the norm in most practical applications, and that solutions based on traditional search methods are not effective in this context. He then presents a solution based on dynamic taxonomies, a knowledge management model that effectively guides users to reach their goal while giving them total freedom in exploring the information base. Applications, benefits, and current research are discussed.
    Date
    22. 7.2006 17:56:22
  12. Efthimiadis, E.N.: End-users' understanding of thesaural knowledge structures in interactive query expansion (1994) 0.02
    0.018260393 = product of:
      0.06391137 = sum of:
        0.04150931 = weight(_text_:management in 5693) [ClassicSimilarity], result of:
          0.04150931 = score(doc=5693,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.29792285 = fieldWeight in 5693, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.0625 = fieldNorm(doc=5693)
        0.02240206 = product of:
          0.04480412 = sum of:
            0.04480412 = weight(_text_:22 in 5693) [ClassicSimilarity], result of:
              0.04480412 = score(doc=5693,freq=2.0), product of:
                0.14475311 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.041336425 = queryNorm
                0.30952093 = fieldWeight in 5693, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5693)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Date
    30. 3.2001 13:35:22
    Source
    Knowledge organization and quality management: Proc. of the 3rd International ISKO Conference, 20-24 June 1994, Copenhagen, Denmark. Ed.: H. Albrechtsen et al
  13. Gao, J.; Zhang, J.: Clustered SVD strategies in latent semantic indexing (2005) 0.02
    0.017649153 = product of:
      0.061772034 = sum of:
        0.036320645 = weight(_text_:management in 1166) [ClassicSimilarity], result of:
          0.036320645 = score(doc=1166,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.2606825 = fieldWeight in 1166, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.0546875 = fieldNorm(doc=1166)
        0.02545139 = product of:
          0.05090278 = sum of:
            0.05090278 = weight(_text_:studies in 1166) [ClassicSimilarity], result of:
              0.05090278 = score(doc=1166,freq=2.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.30860704 = fieldWeight in 1166, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1166)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    The text retrieval method using latent semantic indexing (LSI) technique with truncated singular value decomposition (SVD) has been intensively studied in recent years. The SVD reduces the noise contained in the original representation of the term-document matrix and improves the information retrieval accuracy. Recent studies indicate that SVD is mostly useful for small homogeneous data collections. For large inhomogeneous datasets, the performance of the SVD based text retrieval technique may deteriorate. We propose to partition a large inhomogeneous dataset into several smaller ones with clustered structure, on which we apply the truncated SVD. Our experimental results show that the clustered SVD strategies may enhance the retrieval accuracy and reduce the computing and storage costs.
    Source
    Information processing and management. 41(2005) no.5, S.1051-1064
  14. Shiri, A.A.; Revie, C.: Query expansion behavior within a thesaurus-enhanced search environment : a user-centered evaluation (2006) 0.02
    0.016611008 = product of:
      0.058138527 = sum of:
        0.04413724 = weight(_text_:case in 56) [ClassicSimilarity], result of:
          0.04413724 = score(doc=56,freq=2.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.24286987 = fieldWeight in 56, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.0390625 = fieldNorm(doc=56)
        0.0140012875 = product of:
          0.028002575 = sum of:
            0.028002575 = weight(_text_:22 in 56) [ClassicSimilarity], result of:
              0.028002575 = score(doc=56,freq=2.0), product of:
                0.14475311 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.041336425 = queryNorm
                0.19345059 = fieldWeight in 56, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=56)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    The study reported here investigated the query expansion behavior of end-users interacting with a thesaurus-enhanced search system on the Web. Two groups, namely academic staff and postgraduate students, were recruited into this study. Data were collected from 90 searches performed by 30 users using the OVID interface to the CAB abstracts database. Data-gathering techniques included questionnaires, screen capturing software, and interviews. The results presented here relate to issues of search-topic and search-term characteristics, number and types of expanded queries, usefulness of thesaurus terms, and behavioral differences between academic staff and postgraduate students in their interaction. The key conclusions drawn were that (a) academic staff chose more narrow and synonymous terms than did postgraduate students, who generally selected broader and related terms; (b) topic complexity affected users' interaction with the thesaurus in that complex topics required more query expansion and search term selection; (c) users' prior topic-search experience appeared to have a significant effect on their selection and evaluation of thesaurus terms; (d) in 50% of the searches where additional terms were suggested from the thesaurus, users stated that they had not been aware of the terms at the beginning of the search; this observation was particularly noticeable in the case of postgraduate students.
    Date
    22. 7.2006 16:32:43
  15. Cool, C.; Spink, A.: Issues of context in information retrieval (IR) : an introduction to the special issue (2002) 0.02
    0.015127847 = product of:
      0.05294746 = sum of:
        0.031131983 = weight(_text_:management in 2587) [ClassicSimilarity], result of:
          0.031131983 = score(doc=2587,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.22344214 = fieldWeight in 2587, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.046875 = fieldNorm(doc=2587)
        0.021815477 = product of:
          0.043630954 = sum of:
            0.043630954 = weight(_text_:studies in 2587) [ClassicSimilarity], result of:
              0.043630954 = score(doc=2587,freq=2.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.26452032 = fieldWeight in 2587, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2587)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    The subject of context has received a great deal of attention in the information retrieval (IR) literature over the past decade, primarily in studies of information seeking and IR interactions. Recently, attention to context in IR has expanded to address new problems in new environments. In this paper we outline five overlapping dimensions of context which we believe to be important constituent elements and we discuss how they are related to different issues in IR research. The papers in this special issue are summarized with respect to how they represent work that is being conducted within these dimensions of context. We conclude with future areas of research which are needed in order to fully understand the multidimensional nature of context in IR.
    Source
    Information processing and management. 38(2002) no.5, S.605-611
  16. Looking for information : a survey on research on information seeking, needs, and behavior (2012) 0.01
    0.01261064 = product of:
      0.08827448 = sum of:
        0.08827448 = weight(_text_:case in 3802) [ClassicSimilarity], result of:
          0.08827448 = score(doc=3802,freq=2.0), product of:
            0.18173204 = queryWeight, product of:
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.041336425 = queryNorm
            0.48573974 = fieldWeight in 3802, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.3964143 = idf(docFreq=1480, maxDocs=44218)
              0.078125 = fieldNorm(doc=3802)
      0.14285715 = coord(1/7)
    
    Editor
    Case, D.O.
  17. Olmos, R.; Jorge-Botana, G.; Luzón, J.M.; Martín-Cordero, J.I.; León, J.A.: Transforming LSA space dimensions into a rubric for an automatic assessment and feedback system (2016) 0.01
    0.012606538 = product of:
      0.044122882 = sum of:
        0.025943318 = weight(_text_:management in 2878) [ClassicSimilarity], result of:
          0.025943318 = score(doc=2878,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.18620178 = fieldWeight in 2878, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2878)
        0.018179566 = product of:
          0.03635913 = sum of:
            0.03635913 = weight(_text_:studies in 2878) [ClassicSimilarity], result of:
              0.03635913 = score(doc=2878,freq=2.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.22043361 = fieldWeight in 2878, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2878)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    The purpose of this article is to validate, through two empirical studies, a new method for automatic evaluation of written texts, called Inbuilt Rubric, based on the Latent Semantic Analysis (LSA) technique, which constitutes an innovative and distinct turn with respect to LSA application so far. In the first empirical study, evidence of the validity of the method to identify and evaluate the conceptual axes of a text in a sample of 78 summaries by secondary school students is sought. Results show that the proposed method has a significantly higher degree of reliability than classic LSA methods of text evaluation, and displays very high sensitivity to identify which conceptual axes are included or not in each summary. A second study evaluates the method's capacity to interact and provide feedback about quality in a real online system on a sample of 924 discursive texts written by university students. Results show that students improved the quality of their written texts using this system, and also rated the experience very highly. The final conclusion is that this new method opens a very interesting way regarding the role of automatic assessors in the identification of presence/absence and quality of elaboration of relevant conceptual information in texts written by students with lower time costs than the usual LSA-based methods.
    Source
    Information processing and management. 52(2016) no.3, S.359-373
  18. Roy, R.S.; Agarwal, S.; Ganguly, N.; Choudhury, M.: Syntactic complexity of Web search queries through the lenses of language models, networks and users (2016) 0.01
    0.012606538 = product of:
      0.044122882 = sum of:
        0.025943318 = weight(_text_:management in 3188) [ClassicSimilarity], result of:
          0.025943318 = score(doc=3188,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.18620178 = fieldWeight in 3188, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3188)
        0.018179566 = product of:
          0.03635913 = sum of:
            0.03635913 = weight(_text_:studies in 3188) [ClassicSimilarity], result of:
              0.03635913 = score(doc=3188,freq=2.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.22043361 = fieldWeight in 3188, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3188)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    Across the world, millions of users interact with search engines every day to satisfy their information needs. As the Web grows bigger over time, such information needs, manifested through user search queries, also become more complex. However, there has been no systematic study that quantifies the structural complexity of Web search queries. In this research, we make an attempt towards understanding and characterizing the syntactic complexity of search queries using a multi-pronged approach. We use traditional statistical language modeling techniques to quantify and compare the perplexity of queries with natural language (NL). We then use complex network analysis for a comparative analysis of the topological properties of queries issued by real Web users and those generated by statistical models. Finally, we conduct experiments to study whether search engine users are able to identify real queries, when presented along with model-generated ones. The three complementary studies show that the syntactic structure of Web queries is more complex than what n-grams can capture, but simpler than NL. Queries, thus, seem to represent an intermediate stage between syntactic and non-syntactic communication.
    Source
    Information processing and management. 52(2016) no.5, S.923-948
  19. Hazrina, S.; Sharef, N.M.; Ibrahim, H.; Murad, M.A.A.; Noah, S.A.M.: Review on the advancements of disambiguation in semantic question answering system (2017) 0.01
    0.011806425 = product of:
      0.041322485 = sum of:
        0.020754656 = weight(_text_:management in 3292) [ClassicSimilarity], result of:
          0.020754656 = score(doc=3292,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.14896142 = fieldWeight in 3292, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.03125 = fieldNorm(doc=3292)
        0.020567829 = product of:
          0.041135658 = sum of:
            0.041135658 = weight(_text_:studies in 3292) [ClassicSimilarity], result of:
              0.041135658 = score(doc=3292,freq=4.0), product of:
                0.16494368 = queryWeight, product of:
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.041336425 = queryNorm
                0.24939215 = fieldWeight in 3292, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9902744 = idf(docFreq=2222, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3292)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Abstract
    Ambiguity is a potential problem in any semantic question answering (SQA) system due to the nature of idiosyncrasy in composing natural language (NL) question and semantic resources. Thus, disambiguation of SQA systems is a field of ongoing research. Ambiguity occurs in SQA because a word or a sentence can have more than one meaning or multiple words in the same language can share the same meaning. Therefore, an SQA system needs disambiguation solutions to select the correct meaning when the linguistic triples matched with multiple KB concepts, and enumerate similar words especially when linguistic triples do not match with any KB concept. The latest development in this field is a solution for SQA systems that is able to process a complex NL question while accessing open-domain data from linked open data (LOD). The contributions in this paper include (1) formulating an SQA conceptual framework based on an in-depth study of existing SQA processes; (2) identifying the ambiguity types, specifically in English based on an interdisciplinary literature review; (3) highlighting the ambiguity types that had been resolved by the previous SQA studies; and (4) analysing the results of the existing SQA disambiguation solutions, the complexity of NL question processing, and the complexity of data retrieval from KB(s) or LOD. The results of this review demonstrated that out of thirteen types of ambiguity identified in the literature, only six types had been successfully resolved by the previous studies. Efforts to improve the disambiguation are in progress for the remaining unresolved ambiguity types to improve the accuracy of the formulated answers by the SQA system. The remaining ambiguity types are potentially resolved in the identified SQA process based on ambiguity scenarios elaborated in this paper. The results of this review also demonstrated that most existing research on SQA systems have treated the processing of the NL question complexity separate from the processing of the KB structure complexity.
    Source
    Information processing and management. 53(2017) no.1, S.52-69
  20. Efthimiadis, E.N.: User choices : a new yardstick for the evaluation of ranking algorithms for interactive query expansion (1995) 0.01
    0.011412744 = product of:
      0.039944604 = sum of:
        0.025943318 = weight(_text_:management in 5697) [ClassicSimilarity], result of:
          0.025943318 = score(doc=5697,freq=2.0), product of:
            0.13932906 = queryWeight, product of:
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.041336425 = queryNorm
            0.18620178 = fieldWeight in 5697, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.3706124 = idf(docFreq=4130, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5697)
        0.0140012875 = product of:
          0.028002575 = sum of:
            0.028002575 = weight(_text_:22 in 5697) [ClassicSimilarity], result of:
              0.028002575 = score(doc=5697,freq=2.0), product of:
                0.14475311 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.041336425 = queryNorm
                0.19345059 = fieldWeight in 5697, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5697)
          0.5 = coord(1/2)
      0.2857143 = coord(2/7)
    
    Date
    22. 2.1996 13:14:10
    Source
    Information processing and management. 31(1995) no.4, S.605-620

Years

Languages

  • e 99
  • d 6
  • More… Less…

Types

  • a 94
  • el 10
  • m 7
  • r 1
  • s 1
  • More… Less…