Search (65 results, page 1 of 4)

  • × theme_ss:"Suchtaktik"
  • × type_ss:"a"
  • × year_i:[2010 TO 2020}
  1. Hopkins, M.E.; Zavalina, O.L.: Evaluating physicians' serendipitous knowledge discovery in online discovery systems : a new approach (2019) 0.06
    0.06274393 = product of:
      0.0941159 = sum of:
        0.026472148 = weight(_text_:information in 5842) [ClassicSimilarity], result of:
          0.026472148 = score(doc=5842,freq=18.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.2909321 = fieldWeight in 5842, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5842)
        0.06764375 = sum of:
          0.032530803 = weight(_text_:management in 5842) [ClassicSimilarity], result of:
            0.032530803 = score(doc=5842,freq=2.0), product of:
              0.17470726 = queryWeight, product of:
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.0518325 = queryNorm
              0.18620178 = fieldWeight in 5842, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5842)
          0.035112944 = weight(_text_:22 in 5842) [ClassicSimilarity], result of:
            0.035112944 = score(doc=5842,freq=2.0), product of:
              0.18150859 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0518325 = queryNorm
              0.19345059 = fieldWeight in 5842, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5842)
      0.6666667 = coord(2/3)
    
    Abstract
    Purpose A new approach to investigate serendipitous knowledge discovery (SKD) of health information is developed and tested to evaluate the information flow-serendipitous knowledge discovery (IF-SKD) model. The purpose of this paper is to determine the degree to which IF-SKD reflects physicians' information behaviour in a clinical setting and explore how the information system, Spark, designed to support physicians' SKD, meets its goals. Design/methodology/approach The proposed pre-experimental study design employs an adapted version of the McCay-Peet's (2013) and McCay-Peet et al.'s (2015) serendipitous digital environment (SDE) questionnaire research tool to address the complexity associated with defining the way in which SKD is understood and applied in system design. To test the IF-SKD model, the new data analysis approach combining confirmatory factor analysis, data imputation and Monte Carlo simulations was developed. Findings The piloting of the proposed novel analysis approach demonstrated that small sample information behaviour survey data can be meaningfully examined using a confirmatory factor analysis technique. Research limitations/implications This method allows to improve the reliability in measuring SKD and the generalisability of findings. Originality/value This paper makes an original contribution to developing and refining methods and tools of research into information-system-supported serendipitous discovery of information by health providers.
    Date
    20. 1.2015 18:30:22
    Footnote
    Beitrag in einem Special Issue: Innovative Methods in Health Information Behaviour Research.
    Source
    Aslib journal of information management. 71(2019) no.6, S.755-772
  2. Monchaux, S.; Amadieu, F.; Chevalier, A.; Mariné, C.: Query strategies during information searching : effects of prior domain knowledge and complexity of the information problems to be solved (2015) 0.06
    0.059505444 = product of:
      0.089258164 = sum of:
        0.02161442 = weight(_text_:information in 2680) [ClassicSimilarity], result of:
          0.02161442 = score(doc=2680,freq=12.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.23754507 = fieldWeight in 2680, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2680)
        0.06764375 = sum of:
          0.032530803 = weight(_text_:management in 2680) [ClassicSimilarity], result of:
            0.032530803 = score(doc=2680,freq=2.0), product of:
              0.17470726 = queryWeight, product of:
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.0518325 = queryNorm
              0.18620178 = fieldWeight in 2680, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2680)
          0.035112944 = weight(_text_:22 in 2680) [ClassicSimilarity], result of:
            0.035112944 = score(doc=2680,freq=2.0), product of:
              0.18150859 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0518325 = queryNorm
              0.19345059 = fieldWeight in 2680, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2680)
      0.6666667 = coord(2/3)
    
    Abstract
    This study addresses the impact of domain expertise (i.e. of prior knowledge of the domain) on the performance and query strategies used by users while searching for information. Twenty-four experts (psychology students) and 24 non-experts (students from other disciplines) had to search for psychology information from the Universalis website in order to perform six information problems of varying complexity: two simple problems (the keywords required to complete the task were provided in the problem statement), two more difficult problems (the keywords required had to be inferred) and two impossible problems (no answer was provided by the website). The results showed that participants with prior knowledge in the domain (experts in psychology) performed better (i.e. reached more correct answers after shorter search times) than non-experts. This difference was stronger as the complexity of the problems increased. This study also showed that experts and non-experts displayed different query strategies. Experts reformulated the impossible problems more often than non-experts, because they produced new queries with psychology-related keywords. The participants rarely used thematic category tool and when they did so this did not enhance their performance.
    Date
    25. 1.2016 18:46:22
    Source
    Information processing and management. 51(2015) no.5, S.557-569
  3. Sachse, J.: ¬The influence of snippet length on user behavior in mobile web search (2019) 0.06
    0.05528497 = product of:
      0.08292745 = sum of:
        0.015283704 = weight(_text_:information in 5493) [ClassicSimilarity], result of:
          0.015283704 = score(doc=5493,freq=6.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.16796975 = fieldWeight in 5493, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5493)
        0.06764375 = sum of:
          0.032530803 = weight(_text_:management in 5493) [ClassicSimilarity], result of:
            0.032530803 = score(doc=5493,freq=2.0), product of:
              0.17470726 = queryWeight, product of:
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.0518325 = queryNorm
              0.18620178 = fieldWeight in 5493, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5493)
          0.035112944 = weight(_text_:22 in 5493) [ClassicSimilarity], result of:
            0.035112944 = score(doc=5493,freq=2.0), product of:
              0.18150859 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0518325 = queryNorm
              0.19345059 = fieldWeight in 5493, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=5493)
      0.6666667 = coord(2/3)
    
    Abstract
    Purpose Web search is more and more moving into mobile contexts. However, screen size of mobile devices is limited and search engine result pages face a trade-off between offering informative snippets and optimal use of space. One factor clearly influencing this trade-off is snippet length. The purpose of this paper is to find out what snippet size to use in mobile web search. Design/methodology/approach For this purpose, an eye-tracking experiment was conducted showing participants search interfaces with snippets of one, three or five lines on a mobile device to analyze 17 dependent variables. In total, 31 participants took part in the study. Each of the participants solved informational and navigational tasks. Findings Results indicate a strong influence of page fold on scrolling behavior and attention distribution across search results. Regardless of query type, short snippets seem to provide too little information about the result, so that search performance and subjective measures are negatively affected. Long snippets of five lines lead to better performance than medium snippets for navigational queries, but to worse performance for informational queries. Originality/value Although space in mobile search is limited, this study shows that longer snippets improve usability and user experience. It further emphasizes that page fold plays a stronger role in mobile than in desktop search for attention distribution.
    Date
    20. 1.2015 18:30:22
    Footnote
    Beitag in einem Special Issue: Information Science in the German-speaking Countries
    Source
    Aslib journal of information management. 71(2019) no.3, S.325-343
  4. Renugadevi, S.; Geetha, T.V.; Gayathiri, R.L.; Prathyusha, S.; Kaviya, T.: Collaborative search using an implicitly formed academic network (2014) 0.04
    0.040782824 = product of:
      0.061174236 = sum of:
        0.00705924 = weight(_text_:information in 1628) [ClassicSimilarity], result of:
          0.00705924 = score(doc=1628,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.0775819 = fieldWeight in 1628, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=1628)
        0.054114997 = sum of:
          0.026024643 = weight(_text_:management in 1628) [ClassicSimilarity], result of:
            0.026024643 = score(doc=1628,freq=2.0), product of:
              0.17470726 = queryWeight, product of:
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.0518325 = queryNorm
              0.14896142 = fieldWeight in 1628, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.3706124 = idf(docFreq=4130, maxDocs=44218)
                0.03125 = fieldNorm(doc=1628)
          0.028090354 = weight(_text_:22 in 1628) [ClassicSimilarity], result of:
            0.028090354 = score(doc=1628,freq=2.0), product of:
              0.18150859 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0518325 = queryNorm
              0.15476047 = fieldWeight in 1628, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=1628)
      0.6666667 = coord(2/3)
    
    Date
    20. 1.2015 18:30:22
    Source
    Aslib journal of information management. 66(2014) no.5, S.537-552
  5. Saastamoinen, M.; Järvelin, K.: Search task features in work tasks of varying types and complexity (2017) 0.03
    0.03272217 = product of:
      0.049083255 = sum of:
        0.028015489 = weight(_text_:information in 3589) [ClassicSimilarity], result of:
          0.028015489 = score(doc=3589,freq=14.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.3078936 = fieldWeight in 3589, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3589)
        0.021067765 = product of:
          0.04213553 = sum of:
            0.04213553 = weight(_text_:22 in 3589) [ClassicSimilarity], result of:
              0.04213553 = score(doc=3589,freq=2.0), product of:
                0.18150859 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0518325 = queryNorm
                0.23214069 = fieldWeight in 3589, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3589)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Information searching in practice seldom is an end in itself. In work, work task (WT) performance forms the context, which information searching should serve. Therefore, information retrieval (IR) systems development/evaluation should take the WT context into account. The present paper analyzes how WT features: task complexity and task types, affect information searching in authentic work: the types of information needs, search processes, and search media. We collected data on 22 information professionals in authentic work situations in three organization types: city administration, universities, and companies. The data comprise 286 WTs and 420 search tasks (STs). The data include transaction logs, video recordings, daily questionnaires, interviews. and observation. The data were analyzed quantitatively. Even if the participants used a range of search media, most STs were simple throughout the data, and up to 42% of WTs did not include searching. WT's effects on STs are not straightforward: different WT types react differently to WT complexity. Due to the simplicity of authentic searching, the WT/ST types in interactive IR experiments should be reconsidered.
    Source
    Journal of the Association for Information Science and Technology. 68(2017) no.5, S.1111-1123
  6. Pontis, S.; Blandford, A.; Greifeneder, E.; Attalla, H.; Neal, D.: Keeping up to date : an academic researcher's information journey (2017) 0.03
    0.031215023 = product of:
      0.046822533 = sum of:
        0.029266061 = weight(_text_:information in 3340) [ClassicSimilarity], result of:
          0.029266061 = score(doc=3340,freq=22.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.32163754 = fieldWeight in 3340, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3340)
        0.017556472 = product of:
          0.035112944 = sum of:
            0.035112944 = weight(_text_:22 in 3340) [ClassicSimilarity], result of:
              0.035112944 = score(doc=3340,freq=2.0), product of:
                0.18150859 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0518325 = queryNorm
                0.19345059 = fieldWeight in 3340, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3340)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Keeping up to date with research developments is a central activity of academic researchers, but researchers face difficulties in managing the rapid growth of available scientific information. This study examined how researchers stay up to date, using the information journey model as a framework for analysis and investigating which dimensions influence information behaviors. We designed a 2-round study involving semistructured interviews and prototype testing with 61 researchers with 3 levels of seniority (PhD student to professor). Data were analyzed following a semistructured qualitative approach. Five key dimensions that influence information behaviors were identified: level of seniority, information sources, state of the project, level of familiarity, and how well defined the relevant community is. These dimensions are interrelated and their values determine the flow of the information journey. Across all levels of professional expertise, researchers used similar hard (formal) sources to access content, while soft (interpersonal) sources were used to filter information. An important "pain point" that future information tools should address is helping researchers filter information at the point of need.
    Source
    Journal of the Association for Information Science and Technology. 68(2017) no.1, S.22-35
  7. Sanchiza, M.; Chinb, J.; Chevaliera, A.; Fuc, W.T.; Amadieua, F.; Hed, J.: Searching for information on the web : impact of cognitive aging, prior domain knowledge and complexity of the search problems (2017) 0.03
    0.028797261 = product of:
      0.043195892 = sum of:
        0.02367741 = weight(_text_:information in 3294) [ClassicSimilarity], result of:
          0.02367741 = score(doc=3294,freq=10.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.2602176 = fieldWeight in 3294, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3294)
        0.019518482 = product of:
          0.039036963 = sum of:
            0.039036963 = weight(_text_:management in 3294) [ClassicSimilarity], result of:
              0.039036963 = score(doc=3294,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.22344214 = fieldWeight in 3294, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3294)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This study focuses on the impact of age, prior domain knowledge and cognitive abilities on performance, query production and navigation strategies during information searching. Twenty older adults and nineteen young adults had to answer 12 information search problems of varying nature within two domain knowledge: health and manga. In each domain, participants had to perform two simple fact-finding problems (keywords provided and answer directly accessible on the search engine results page), two difficult fact-finding problems (keywords had to be inferred) and two open-ended information search problems (multiple answers possible and navigation necessary). Results showed that prior domain knowledge helped older adults improve navigation (i.e. reduced the number of webpages visited and thus decreased the feeling of disorientation), query production and reformulation (i.e. they formulated semantically more specific queries, and they inferred a greater number of new keywords).
    Source
    Information processing and management. 53(2017) no.1, S.281-294
  8. Vuong, T.; Saastamoinen, M.; Jacucci, G.; Ruotsalo, T.: Understanding user behavior in naturalistic information search tasks (2019) 0.03
    0.028491702 = product of:
      0.04273755 = sum of:
        0.026472148 = weight(_text_:information in 5419) [ClassicSimilarity], result of:
          0.026472148 = score(doc=5419,freq=18.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.2909321 = fieldWeight in 5419, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5419)
        0.016265402 = product of:
          0.032530803 = sum of:
            0.032530803 = weight(_text_:management in 5419) [ClassicSimilarity], result of:
              0.032530803 = score(doc=5419,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.18620178 = fieldWeight in 5419, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5419)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Understanding users' search behavior has largely relied on the information available from search engine logs, which provide limited information about the contextual factors affecting users' behavior. Consequently, questions such as how users' intentions, task goals, and substances of the users' tasks affect search behavior, as well as what triggers information needs, remain largely unanswered. We report an experiment in which naturalistic information search behavior was captured by analyzing 24/7 continuous recordings of information on participants' computer screens. Written task diaries describing the participants' tasks were collected and used as real-life task contexts for further categorization. All search tasks were extracted and classified under various task categories according to users' intentions, task goals, and substances of the tasks. We investigated the effect of different task categories on three behavioral factors: search efforts, content-triggers, and application context. Our results suggest four findings: (i) Search activity is integrally associated with the users' creative processes. The content users have seen prior to searching more often triggers search, and is used as a query, within creative tasks. (ii) Searching within intellectual and creative tasks is more time-intensive, while search activity occurring as a part of daily routine tasks is associated with more frequent searching within a search task. (iii) Searching is more often induced from utility applications in tasks demanding a degree of intellectual effort. (iv) Users' leisure information-seeking activity is occurring inherently within social media services or comes from social communication platforms. The implications of our findings for information access and management systems are discussed.
    Source
    Journal of the Association for Information Science and Technology. 70(2019) no.11, S.1248-1261
  9. Tamine, L.; Chouquet, C.: On the impact of domain expertise on query formulation, relevance assessment and retrieval performance in clinical settings (2017) 0.02
    0.02399772 = product of:
      0.03599658 = sum of:
        0.019731175 = weight(_text_:information in 3290) [ClassicSimilarity], result of:
          0.019731175 = score(doc=3290,freq=10.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.21684799 = fieldWeight in 3290, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3290)
        0.016265402 = product of:
          0.032530803 = sum of:
            0.032530803 = weight(_text_:management in 3290) [ClassicSimilarity], result of:
              0.032530803 = score(doc=3290,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.18620178 = fieldWeight in 3290, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3290)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The large volumes of medical information available on the web may provide answers for a wide range of users attempting to solve health-related problems. While experts generally utilize reliable resources for diagnosis search and professional development, novices utilize different (social) web resources to obtain information that helps them manage their health or the health of people who they care for. A diverse number of related search topics address clinical diagnosis, advice searching, information sharing, connecting with experts, etc. This paper focuses on the extent to which expertise can impact clinical query formulation, document relevance assessment and retrieval performance in the context of tailoring retrieval models and systems to experts vs. non-experts. The results show that medical domain expertise 1) plays an important role in the lexical representations of information needs; 2) significantly influences the perception of relevance even among users with similar levels of expertise and 3) reinforces the idea that a single ground truth does not exist, thereby leading to the variability of system rankings with respect to the level of user's expertise. The findings of this study presents opportunities for the design of personalized health-related IR systems, but also for providing insights about the evaluation of such systems.
    Source
    Information processing and management. 53(2017) no.2, S.332-350
  10. Russell-Rose, T.; Chamberlain, J.; Azzopardi, L.: Information retrieval in the workplace : a comparison of professional search practices (2018) 0.02
    0.02399772 = product of:
      0.03599658 = sum of:
        0.019731175 = weight(_text_:information in 5048) [ClassicSimilarity], result of:
          0.019731175 = score(doc=5048,freq=10.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.21684799 = fieldWeight in 5048, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5048)
        0.016265402 = product of:
          0.032530803 = sum of:
            0.032530803 = weight(_text_:management in 5048) [ClassicSimilarity], result of:
              0.032530803 = score(doc=5048,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.18620178 = fieldWeight in 5048, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5048)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Legal researchers, recruitment professionals, healthcare information professionals, and patent analysts all undertake work tasks where search forms a core part of their duties. In these instances, the search task is often complex and time-consuming and requires specialist expertise to identify relevant documents and insights within large domain-specific repositories and collections. Several studies have been made investigating the search practices of professionals such as these, but few have attempted to directly compare their professional practices and so it remains unclear to what extent insights and approaches from one domain can be applied to another. In this paper we describe the results of a survey of a purposive sample of 108 legal researchers, 64 recruitment professionals and 107 healthcare information professionals. Their responses are compared with results from a previous survey of 81 patent analysts. The survey investigated their search practices and preferences, the types of functionality they value, and their requirements for future information retrieval systems. The results reveal that these professions share many fundamental needs and face similar challenges. In particular a continuing preference to formulate queries as Boolean expressions, the need to manage, organise and re-use search strategies and results and an ambivalence toward the use of relevance ranking. The results stress the importance of recall and coverage for the healthcare and patent professionals, while precision and recency were more important to the legal and recruitment professionals. The results also highlight the need to ensure that search systems give confidence to the professional searcher and so trust, explainability and accountability remains a significant challenge when developing such systems. The findings suggest that translational research between the different areas could benefit professionals across domains.
    Source
    Information processing and management. 54(2018) no.6, S.1042-1057
  11. Barrio, P.; Gravano, L.: Sampling strategies for information extraction over the deep web (2017) 0.02
    0.023557067 = product of:
      0.0353356 = sum of:
        0.022323277 = weight(_text_:information in 3412) [ClassicSimilarity], result of:
          0.022323277 = score(doc=3412,freq=20.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.2453355 = fieldWeight in 3412, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=3412)
        0.013012322 = product of:
          0.026024643 = sum of:
            0.026024643 = weight(_text_:management in 3412) [ClassicSimilarity], result of:
              0.026024643 = score(doc=3412,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.14896142 = fieldWeight in 3412, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3412)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Information extraction systems discover structured information in natural language text. Having information in structured form enables much richer querying and data mining than possible over the natural language text. However, information extraction is a computationally expensive task, and hence improving the efficiency of the extraction process over large text collections is of critical interest. In this paper, we focus on an especially valuable family of text collections, namely, the so-called deep-web text collections, whose contents are not crawlable and are only available via querying. Important steps for efficient information extraction over deep-web text collections (e.g., selecting the collections on which to focus the extraction effort, based on their contents; or learning which documents within these collections-and in which order-to process, based on their words and phrases) require having a representative document sample from each collection. These document samples have to be collected by querying the deep-web text collections, an expensive process that renders impractical the existing sampling approaches developed for other data scenarios. In this paper, we systematically study the space of query-based document sampling techniques for information extraction over the deep web. Specifically, we consider (i) alternative query execution schedules, which vary on how they account for the query effectiveness, and (ii) alternative document retrieval and processing schedules, which vary on how they distribute the extraction effort over documents. We report the results of the first large-scale experimental evaluation of sampling techniques for information extraction over the deep web. Our results show the merits and limitations of the alternative query execution and document retrieval and processing strategies, and provide a roadmap for addressing this critically important building block for efficient, scalable information extraction.
    Source
    Information processing and management. 53(2017) no.2, S.309-331
  12. Aloteibi, S.; Sanderson, M.: Analyzing geographic query reformulation : an exploratory study (2014) 0.02
    0.023469714 = product of:
      0.03520457 = sum of:
        0.017648099 = weight(_text_:information in 1177) [ClassicSimilarity], result of:
          0.017648099 = score(doc=1177,freq=8.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.19395474 = fieldWeight in 1177, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1177)
        0.017556472 = product of:
          0.035112944 = sum of:
            0.035112944 = weight(_text_:22 in 1177) [ClassicSimilarity], result of:
              0.035112944 = score(doc=1177,freq=2.0), product of:
                0.18150859 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0518325 = queryNorm
                0.19345059 = fieldWeight in 1177, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1177)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    Search engine users typically engage in multiquery sessions in their quest to fulfill their information needs. Despite a plethora of research findings suggesting that a significant group of users look for information within a specific geographical scope, existing reformulation studies lack a focused analysis of how users reformulate geographic queries. This study comprehensively investigates the ways in which users reformulate such needs in an attempt to fill this gap in the literature. Reformulated sessions were sampled from a query log of a major search engine to extract 2,400 entries that were manually inspected to filter geo sessions. This filter identified 471 search sessions that included geographical intent, and these sessions were analyzed quantitatively and qualitatively. The results revealed that one in five of the users who reformulated their queries were looking for geographically related information. They reformulated their queries by changing the content of the query rather than the structure. Users were not following a unified sequence of modifications and instead performed a single reformulation action. However, in some cases it was possible to anticipate their next move. A number of tasks in geo modifications were identified, including standard, multi-needs, multi-places, and hybrid approaches. The research concludes that it is important to specialize query reformulation studies to focus on particular query types rather than generically analyzing them, as it is apparent that geographic queries have their special reformulation characteristics.
    Date
    26. 1.2014 18:48:22
    Source
    Journal of the Association for Information Science and Technology. 65(2014) no.1, S.13-24
  13. Wu, I.-C.; Vakkari, P.: Effects of subject-oriented visualization tools on search by novices and intermediates (2018) 0.02
    0.023469714 = product of:
      0.03520457 = sum of:
        0.017648099 = weight(_text_:information in 4573) [ClassicSimilarity], result of:
          0.017648099 = score(doc=4573,freq=8.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.19395474 = fieldWeight in 4573, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4573)
        0.017556472 = product of:
          0.035112944 = sum of:
            0.035112944 = weight(_text_:22 in 4573) [ClassicSimilarity], result of:
              0.035112944 = score(doc=4573,freq=2.0), product of:
                0.18150859 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0518325 = queryNorm
                0.19345059 = fieldWeight in 4573, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4573)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This study explores how user subject knowledge influences search task processes and outcomes, as well as how search behavior is influenced by subject-oriented information visualization (IV) tools. To enable integrated searches, the proposed WikiMap + integrates search functions and IV tools (i.e., a topic network and hierarchical topic tree) and gathers information from Wikipedia pages and Google Search results. To evaluate the effectiveness of the proposed interfaces, we design subject-oriented tasks and adopt extended evaluation measures. We recruited 48 novices and 48 knowledgeable users, that is, intermediates, for the evaluation. Our results show that novices using the proposed interface demonstrate better search performance than intermediates using Wikipedia. We therefore conclude that our tools help close the gap between novices and intermediates in information searches. The results also show that intermediates can take advantage of the search tool by leveraging the IV tools to browse subtopics, and formulate better queries with less effort. We conclude that embedding the IV and the search tools in the interface can result in different search behavior but improved task performance. We provide implications to design search systems to include IV features adapted to user levels of subject knowledge to help them achieve better task performance.
    Date
    9.12.2018 16:22:25
    Source
    Journal of the Association for Information Science and Technology. 69(2018) no.12, S.1428-1445
  14. Abacha, A.B.; Zweigenbaum, P.: MEANS: A medical question-answering system combining NLP techniques and semantic Web technologies (2015) 0.02
    0.022995595 = product of:
      0.03449339 = sum of:
        0.014974909 = weight(_text_:information in 2677) [ClassicSimilarity], result of:
          0.014974909 = score(doc=2677,freq=4.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.16457605 = fieldWeight in 2677, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2677)
        0.019518482 = product of:
          0.039036963 = sum of:
            0.039036963 = weight(_text_:management in 2677) [ClassicSimilarity], result of:
              0.039036963 = score(doc=2677,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.22344214 = fieldWeight in 2677, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2677)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    The Question Answering (QA) task aims to provide precise and quick answers to user questions from a collection of documents or a database. This kind of IR system is sorely needed with the dramatic growth of digital information. In this paper, we address the problem of QA in the medical domain where several specific conditions are met. We propose a semantic approach to QA based on (i) Natural Language Processing techniques, which allow a deep analysis of medical questions and documents and (ii) semantic Web technologies at both representation and interrogation levels. We present our Semantic Question-Answering System, called MEANS and our proposed method for "Answer Search" based on semantic search and query relaxation. We evaluate the overall system performance on real questions and answers extracted from MEDLINE articles. Our experiments show promising results and suggest that a query-relaxation strategy can further improve the overall performance.
    Source
    Information processing and management. 51(2015) no.5, S.570-594
  15. Grau, B.: Finding answers to questions, in text collections or Web, in open domain or specialty domains (2012) 0.02
    0.021032736 = product of:
      0.031549104 = sum of:
        0.015283704 = weight(_text_:information in 107) [ClassicSimilarity], result of:
          0.015283704 = score(doc=107,freq=6.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.16796975 = fieldWeight in 107, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=107)
        0.016265402 = product of:
          0.032530803 = sum of:
            0.032530803 = weight(_text_:management in 107) [ClassicSimilarity], result of:
              0.032530803 = score(doc=107,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.18620178 = fieldWeight in 107, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=107)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This chapter is dedicated to factual question answering, i.e., extracting precise and exact answers to question given in natural language from texts. A question in natural language gives more information than a bag of word query (i.e., a query made of a list of words), and provides clues for finding precise answers. The author first focuses on the presentation of the underlying problems mainly due to the existence of linguistic variations between questions and their answerable pieces of texts for selecting relevant passages and extracting reliable answers. The author first presents how to answer factual question in open domain. The author also presents answering questions in specialty domain as it requires dealing with semi-structured knowledge and specialized terminologies, and can lead to different applications, as information management in corporations for example. Searching answers on the Web constitutes another application frame and introduces specificities linked to Web redundancy or collaborative usage. Besides, the Web is also multilingual, and a challenging problem consists in searching answers in target language documents other than the source language of the question. For all these topics, this chapter presents main approaches and the remaining problems.
    Source
    Next generation search engines: advanced models for information retrieval. Eds.: C. Jouis, u.a
  16. Xie, I.; Joo, S.: Factors affecting the selection of search tactics : tasks, knowledge, process, and systems (2012) 0.02
    0.020071562 = product of:
      0.030107342 = sum of:
        0.01058886 = weight(_text_:information in 2739) [ClassicSimilarity], result of:
          0.01058886 = score(doc=2739,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.116372846 = fieldWeight in 2739, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=2739)
        0.019518482 = product of:
          0.039036963 = sum of:
            0.039036963 = weight(_text_:management in 2739) [ClassicSimilarity], result of:
              0.039036963 = score(doc=2739,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.22344214 = fieldWeight in 2739, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2739)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Information processing and management. 48(2012) no.2, S.254-270
  17. Wildemuth, B.M.; Kelly, D,; Boettcher, E.; Moore, E.; Dimitrova, G.: Examining the impact of domain and cognitive complexity on query formulation and reformulation (2018) 0.02
    0.020071562 = product of:
      0.030107342 = sum of:
        0.01058886 = weight(_text_:information in 5085) [ClassicSimilarity], result of:
          0.01058886 = score(doc=5085,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.116372846 = fieldWeight in 5085, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=5085)
        0.019518482 = product of:
          0.039036963 = sum of:
            0.039036963 = weight(_text_:management in 5085) [ClassicSimilarity], result of:
              0.039036963 = score(doc=5085,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.22344214 = fieldWeight in 5085, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5085)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Information processing and management. 54(2018) no.3, S.433-450
  18. Rieh, S.Y.; Kim, Y.-M.; Markey, K.: Amount of invested mental effort (AIME) in online searching (2012) 0.02
    0.019162996 = product of:
      0.028744493 = sum of:
        0.01247909 = weight(_text_:information in 2726) [ClassicSimilarity], result of:
          0.01247909 = score(doc=2726,freq=4.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.13714671 = fieldWeight in 2726, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2726)
        0.016265402 = product of:
          0.032530803 = sum of:
            0.032530803 = weight(_text_:management in 2726) [ClassicSimilarity], result of:
              0.032530803 = score(doc=2726,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.18620178 = fieldWeight in 2726, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2726)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    This research investigates how people's perceptions of information retrieval (IR) systems, their perceptions of search tasks, and their perceptions of self-efficacy influence the amount of invested mental effort (AIME) they put into using two different IR systems: a Web search engine and a library system. It also explores the impact of mental effort on an end user's search experience. To assess AIME in online searching, two experiments were conducted using these methods: Experiment 1 relied on self-reports and Experiment 2 employed the dual-task technique. In both experiments, data were collected through search transaction logs, a pre-search background questionnaire, a post-search questionnaire and an interview. Important findings are these: (1) subjects invested greater mental effort searching a library system than searching the Web; (2) subjects put little effort into Web searching because of their high sense of self-efficacy in their searching ability and their perception of the easiness of the Web; (3) subjects did not recognize that putting mental effort into searching was something needed to improve the search results; and (4) data collected from multiple sources proved to be effective for assessing mental effort in online searching.
    Source
    Information processing and management. 48(2012) no.6, S.1136-1150
  19. Bilal, D.; Gwizdka, J.: Children's query types and reformulations in Google search (2018) 0.02
    0.019162996 = product of:
      0.028744493 = sum of:
        0.01247909 = weight(_text_:information in 5047) [ClassicSimilarity], result of:
          0.01247909 = score(doc=5047,freq=4.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.13714671 = fieldWeight in 5047, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5047)
        0.016265402 = product of:
          0.032530803 = sum of:
            0.032530803 = weight(_text_:management in 5047) [ClassicSimilarity], result of:
              0.032530803 = score(doc=5047,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.18620178 = fieldWeight in 5047, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5047)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Abstract
    We investigated the searching behaviors of twenty-four children in grades 6, 7, and 8 (ages 11-13) in finding information on three types of search tasks in Google. Children conducted 72 search sessions and issued 150 queries. Children's phrase- and question-like queries combined were much more prevalent than keyword queries (70% vs. 30%, respectively). Fifty two percent of the queries were reformulations (33 sessions). We classified children's query reformulation types into five classes based on the taxonomy by Liu et al. (2010). We found that most query reformulations were by Substitution and Specialization, and that children hardly repeated queries. We categorized children's queries by task facets and examined the way they expressed these facets in their query formulations and reformulations. Oldest children tended to target the general topic of search tasks in their queries most frequently, whereas younger children expressed one of the two facets more often. We assessed children's achieved task outcomes using the search task outcomes measure we developed. Children were mostly more successful on the fact-finding and fully self-generated task and partially successful on the research-oriented task. Query type, reformulation type, achieved task outcomes, and expressing task facets varied by task type and grade level. There was no significant effect of query length in words or of the number of queries issued on search task outcomes. The study findings have implications for human intervention, digital literacy, search task literacy, as well as for system intervention to support children's query formulation and reformulation during interaction with Google.
    Source
    Information processing and management. 54(2018) no.6, S.1022-1041
  20. Lykke, M.; Price, S.; Delcambre, L.: How doctors search : a study of query behaviour and the impact on search results (2012) 0.02
    0.0167263 = product of:
      0.02508945 = sum of:
        0.0088240495 = weight(_text_:information in 2727) [ClassicSimilarity], result of:
          0.0088240495 = score(doc=2727,freq=2.0), product of:
            0.09099081 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0518325 = queryNorm
            0.09697737 = fieldWeight in 2727, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2727)
        0.016265402 = product of:
          0.032530803 = sum of:
            0.032530803 = weight(_text_:management in 2727) [ClassicSimilarity], result of:
              0.032530803 = score(doc=2727,freq=2.0), product of:
                0.17470726 = queryWeight, product of:
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0518325 = queryNorm
                0.18620178 = fieldWeight in 2727, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.3706124 = idf(docFreq=4130, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2727)
          0.5 = coord(1/2)
      0.6666667 = coord(2/3)
    
    Source
    Information processing and management. 48(2012) no.6, S.1151-1170

Languages

  • e 64
  • d 1
  • More… Less…