Search (3346 results, page 2 of 168)

  • × type_ss:"a"
  1. Wiggins, R.: Vendors future : Northern Light - delivering high-quality content to a large Internet audience (1997) 0.07
    0.06521748 = product of:
      0.13043496 = sum of:
        0.13043496 = product of:
          0.26086992 = sum of:
            0.26086992 = weight(_text_:light in 2890) [ClassicSimilarity], result of:
              0.26086992 = score(doc=2890,freq=8.0), product of:
                0.2920221 = queryWeight, product of:
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.050563898 = queryNorm
                0.89332247 = fieldWeight in 2890, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2890)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    A new Web based service, Northern Light, aims to serve large populations of users, by delivering high-quality content on both general and narrow topics. Analyzes the trends that have led to an explosion of access to information on the Internet, but also to difficulties in finding relevant, quality information. Describes the Northern Light search engine which improves naive user searching through its innovative refinement scheme Custom Search folders, but also offers a more sophisticated search syntax for finer control. Searching is free, as is access to many Web Sies, but access to full text articles from a special collection of journals is fee-based. Advocates this free saerch / pay for content payment model for the wider information industry
    Object
    Northern Light
  2. Madison, O.M.A.: Standards in light of new technologies : functional requirements for bibliographic records (1999) 0.07
    0.06521748 = product of:
      0.13043496 = sum of:
        0.13043496 = product of:
          0.26086992 = sum of:
            0.26086992 = weight(_text_:light in 4182) [ClassicSimilarity], result of:
              0.26086992 = score(doc=4182,freq=2.0), product of:
                0.2920221 = queryWeight, product of:
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.050563898 = queryNorm
                0.89332247 = fieldWeight in 4182, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.109375 = fieldNorm(doc=4182)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  3. Northern Light bridges two worlds : innovative search service expands possibilities for ordinary Web users (1998) 0.06
    0.06454857 = product of:
      0.12909713 = sum of:
        0.12909713 = product of:
          0.25819427 = sum of:
            0.25819427 = weight(_text_:light in 2607) [ClassicSimilarity], result of:
              0.25819427 = score(doc=2607,freq=6.0), product of:
                0.2920221 = queryWeight, product of:
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.050563898 = queryNorm
                0.88416 = fieldWeight in 2607, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2607)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Northern Light is an innovative service which creatively blends Web and proprietary online searching. It simultaneously searches the Web and a large, multidisciplinary, full text database, using a relevance system with some clever tweaks. Its risky pricing scheme depends upon users' willingness to pay for proprietary content
    Object
    Northern Light
  4. Jascó, P.: Northern Light (1998) 0.06
    0.06454857 = product of:
      0.12909713 = sum of:
        0.12909713 = product of:
          0.25819427 = sum of:
            0.25819427 = weight(_text_:light in 3310) [ClassicSimilarity], result of:
              0.25819427 = score(doc=3310,freq=6.0), product of:
                0.2920221 = queryWeight, product of:
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.050563898 = queryNorm
                0.88416 = fieldWeight in 3310, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3310)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Northern Light is part WWW search engine and part full text database. The latter is called Special Collection and consists of fulltext articles from 1.800 journals, newswires and other resources. Searching, bibliographic information and summaries are free but with prices per article ranging from $1 to $4 of a monthly subscription for 50 documents from a 880 journal subset. Highlights weaknesses with the software
    Object
    Northern Light
  5. Byström, K.: Information seekers in context : an analysis of the 'doer' in INSU studies (1999) 0.06
    0.06371069 = product of:
      0.12742138 = sum of:
        0.12742138 = sum of:
          0.09316782 = weight(_text_:light in 297) [ClassicSimilarity], result of:
            0.09316782 = score(doc=297,freq=2.0), product of:
              0.2920221 = queryWeight, product of:
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.050563898 = queryNorm
              0.31904373 = fieldWeight in 297, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.0390625 = fieldNorm(doc=297)
          0.034253553 = weight(_text_:22 in 297) [ClassicSimilarity], result of:
            0.034253553 = score(doc=297,freq=2.0), product of:
              0.17706616 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050563898 = queryNorm
              0.19345059 = fieldWeight in 297, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=297)
      0.5 = coord(1/2)
    
    Abstract
    In information needs, seeking and use (INSU) research, individuals have most commonly been perceived as users (e.g., Kuhlthau, 1991; Dervin & Nilan, 1986; Dervin, 1989; Belkin, 1980). The concept user originates from the user of libraries and other information services and information systems. Over the years the scope of the concept has become wider and it is nowadays often understood in the sense of seekers of information (e.g., Wilson, 1981; Marchionini, 1995) and users of information (e.g., Streatfield, 1983). Nevertheless, the concept has remained ambiguous by being on the one hand universal and on the other hand extremely specific. The purpose of this paper is to map and evaluate views on people whose information behaviour has been in one way or another the core of our research area. The goal is to shed some light on various relationships between the different aspects of doers in INSU studies. The paper is inspired by Dervin's (1997) analysis of context where she identified among other themes the nature of subject by contrasting a `transcendental individual' with a `decentered subject', and Talja's (1997) presentation about constituting `information' and `user' from the discourse analytic viewpoint as opposed to the cognitive viewpoint. Instead of the metatheoretical approach applied by Dervin and Talja, a more concrete approach is valid in the present analysis where no direct arguments for or against the underlying metatheories are itemised. The focus is on doers in INSU studies leaving other, even closely-related concepts (i.e., information, information seeking, knowledge etc.), outside the scope of the paper.
    Date
    22. 3.2002 9:55:52
  6. Song, D.; Bruza, P.D.: Towards context sensitive information inference (2003) 0.06
    0.06371069 = product of:
      0.12742138 = sum of:
        0.12742138 = sum of:
          0.09316782 = weight(_text_:light in 1428) [ClassicSimilarity], result of:
            0.09316782 = score(doc=1428,freq=2.0), product of:
              0.2920221 = queryWeight, product of:
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.050563898 = queryNorm
              0.31904373 = fieldWeight in 1428, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1428)
          0.034253553 = weight(_text_:22 in 1428) [ClassicSimilarity], result of:
            0.034253553 = score(doc=1428,freq=2.0), product of:
              0.17706616 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050563898 = queryNorm
              0.19345059 = fieldWeight in 1428, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1428)
      0.5 = coord(1/2)
    
    Abstract
    Humans can make hasty, but generally robust judgements about what a text fragment is, or is not, about. Such judgements are termed information inference. This article furnishes an account of information inference from a psychologistic stance. By drawing an theories from nonclassical logic and applied cognition, an information inference mechanism is proposed that makes inferences via computations of information flow through an approximation of a conceptual space. Within a conceptual space information is represented geometrically. In this article, geometric representations of words are realized as vectors in a high dimensional semantic space, which is automatically constructed from a text corpus. Two approaches were presented for priming vector representations according to context. The first approach uses a concept combination heuristic to adjust the vector representation of a concept in the light of the representation of another concept. The second approach computes a prototypical concept an the basis of exemplar trace texts and moves it in the dimensional space according to the context. Information inference is evaluated by measuring the effectiveness of query models derived by information flow computations. Results show that information flow contributes significantly to query model effectiveness, particularly with respect to precision. Moreover, retrieval effectiveness compares favorably with two probabilistic query models, and another based an semantic association. More generally, this article can be seen as a contribution towards realizing operational systems that mimic text-based human reasoning.
    Date
    22. 3.2003 19:35:46
  7. Dousa, T.M.: ¬The simple and the complex in E. C. Richardson's theory of classification : observations on an early KO model of the relationship between ontology and epistemology (2010) 0.06
    0.06371069 = product of:
      0.12742138 = sum of:
        0.12742138 = sum of:
          0.09316782 = weight(_text_:light in 3509) [ClassicSimilarity], result of:
            0.09316782 = score(doc=3509,freq=2.0), product of:
              0.2920221 = queryWeight, product of:
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.050563898 = queryNorm
              0.31904373 = fieldWeight in 3509, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.0390625 = fieldNorm(doc=3509)
          0.034253553 = weight(_text_:22 in 3509) [ClassicSimilarity], result of:
            0.034253553 = score(doc=3509,freq=2.0), product of:
              0.17706616 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050563898 = queryNorm
              0.19345059 = fieldWeight in 3509, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=3509)
      0.5 = coord(1/2)
    
    Abstract
    In light of ongoing debates about ontological vs. epistemological approaches to knowledge organization (KO), this paper examines E. C. Richardson's treatment of ontology and epistemology in his theory of classification. According to Richardson, there is a natural order of things in the world accessible to human cognition, which may be expressed in two classificatory orders: evolutionary classification, which ranges classes of things from the most simple to the most complex, and logical classification, which ranges classes of things in the inverse order, from the most complex to the most simple. Evolutionary classification reflects ontological order and logical classification reflects epistemological order: both are faces of a single natural order. Such a view requires adherence to a representationalist, or, in Hjorland's (2008) terms, positivist understanding of epistemology, wherein human knowledge faithfully mirrors the structure of the external world. Richardson's harmonization of ontology and epistemology will find little favor among proponents of the currently fashionable pragmatist approach to KO. Nevertheless, it constitutes an early version of what Gnoli (2004) terms a naturalistic epistemology, which, once deepened and clarified, offers the best prospects for an explicit expression of both the ontological and epistemological dimensions of knowledge within a single classification of general scope.
    Pages
    S.15-22
  8. Stvilia, B.; Hinnant, C.C.; Schindler, K.; Worrall, A.; Burnett, G.; Burnett, K.; Kazmer, M.M.; Marty, P.F.: Composition of scientific teams and publication productivity at a national science lab (2011) 0.06
    0.06371069 = product of:
      0.12742138 = sum of:
        0.12742138 = sum of:
          0.09316782 = weight(_text_:light in 4191) [ClassicSimilarity], result of:
            0.09316782 = score(doc=4191,freq=2.0), product of:
              0.2920221 = queryWeight, product of:
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.050563898 = queryNorm
              0.31904373 = fieldWeight in 4191, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.0390625 = fieldNorm(doc=4191)
          0.034253553 = weight(_text_:22 in 4191) [ClassicSimilarity], result of:
            0.034253553 = score(doc=4191,freq=2.0), product of:
              0.17706616 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050563898 = queryNorm
              0.19345059 = fieldWeight in 4191, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=4191)
      0.5 = coord(1/2)
    
    Abstract
    The production of scientific knowledge has evolved from a process of inquiry largely based on the activities of individual scientists to one grounded in the collaborative efforts of specialized research teams. This shift brings to light a new question: how the composition of scientific teams affects their production of knowledge. This study employs data from 1,415 experiments conducted at the National High Magnetic Field Laboratory (NHMFL) between 2005 and 2008 to identify and select a sample of 89 teams and examine whether team diversity and network characteristics affect productivity. The study examines how the diversity of science teams along several variables affects overall team productivity. Results indicate several diversity measures associated with network position and team productivity. Teams with mixed institutional associations were more central to the overall network compared with teams that primarily comprised NHMFL's own scientists. Team cohesion was positively related to productivity. The study indicates that high productivity in teams is associated with high disciplinary diversity and low seniority diversity of team membership. Finally, an increase in the share of senior members negatively affects productivity, and teams with members in central structural positions perform better than other teams.
    Date
    22. 1.2011 13:19:42
  9. Baião Salgado Silva, G.; Lima, G.Â. Borém de Oliveira: Using topic maps in establishing compatibility of semantically structured hypertext contents (2012) 0.06
    0.06371069 = product of:
      0.12742138 = sum of:
        0.12742138 = sum of:
          0.09316782 = weight(_text_:light in 633) [ClassicSimilarity], result of:
            0.09316782 = score(doc=633,freq=2.0), product of:
              0.2920221 = queryWeight, product of:
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.050563898 = queryNorm
              0.31904373 = fieldWeight in 633, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.0390625 = fieldNorm(doc=633)
          0.034253553 = weight(_text_:22 in 633) [ClassicSimilarity], result of:
            0.034253553 = score(doc=633,freq=2.0), product of:
              0.17706616 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050563898 = queryNorm
              0.19345059 = fieldWeight in 633, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=633)
      0.5 = coord(1/2)
    
    Abstract
    Considering the characteristics of hypertext systems and problems such as cognitive overload and the disorientation of users, this project studies subject hypertext documents that have undergone conceptual structuring using facets for content representation and improvement of information retrieval during navigation. The main objective was to assess the possibility of the application of topic map technology for automating the compatibilization process of these structures. For this purpose, two dissertations from the UFMG Information Science Post-Graduation Program were adopted as samples. Both dissertations had been duly analyzed and structured on the MHTX (Hypertextual Map) prototype database. The faceted structures of both dissertations, which had been represented in conceptual maps, were then converted into topic maps. It was then possible to use the merge property of the topic maps to promote the semantic interrelationship between the maps and, consequently, between the hypertextual information resources proper. The merge results were then analyzed in the light of theories dealing with the compatibilization of languages developed within the realm of information technology and librarianship from the 1960s on. The main goals accomplished were: (a) the detailed conceptualization of the merge process of the topic maps, considering the possible compatibilization levels and the applicability of this technology in the integration of faceted structures; and (b) the production of a detailed sequence of steps that may be used in the implementation of topic maps based on faceted structures.
    Date
    22. 2.2013 11:39:23
  10. Lee, J.H.; Price, R.: User experience with commercial music services : an empirical exploration (2016) 0.06
    0.06371069 = product of:
      0.12742138 = sum of:
        0.12742138 = sum of:
          0.09316782 = weight(_text_:light in 2845) [ClassicSimilarity], result of:
            0.09316782 = score(doc=2845,freq=2.0), product of:
              0.2920221 = queryWeight, product of:
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.050563898 = queryNorm
              0.31904373 = fieldWeight in 2845, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2845)
          0.034253553 = weight(_text_:22 in 2845) [ClassicSimilarity], result of:
            0.034253553 = score(doc=2845,freq=2.0), product of:
              0.17706616 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050563898 = queryNorm
              0.19345059 = fieldWeight in 2845, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=2845)
      0.5 = coord(1/2)
    
    Abstract
    The music information retrieval (MIR) community has long understood the role of evaluation as a critical component for successful information retrieval systems. Over the past several years, it has also become evident that user-centered evaluation based on realistic tasks is essential for creating systems that are commercially marketable. Although user-oriented research has been increasing, the MIR field is still lacking in holistic, user-centered approaches to evaluating music services beyond measuring the performance of search or classification algorithms. In light of this need, we conducted a user study exploring how users evaluate their overall experience with existing popular commercial music services, asking about their interactions with the system as well as situational and personal characteristics. In this paper, we present a qualitative heuristic evaluation of commercial music services based on Jakob Nielsen's 10 usability heuristics for user interface design, and also discuss 8 additional criteria that may be used for the holistic evaluation of user experience in MIR systems. Finally, we recommend areas of future user research raised by trends and patterns that surfaced from this user study.
    Date
    17. 3.2016 19:22:15
  11. Ilhan, A.; Fietkiewicz, K.J.: Data privacy-related behavior and concerns of activity tracking technology users from Germany and the USA (2021) 0.06
    0.06371069 = product of:
      0.12742138 = sum of:
        0.12742138 = sum of:
          0.09316782 = weight(_text_:light in 180) [ClassicSimilarity], result of:
            0.09316782 = score(doc=180,freq=2.0), product of:
              0.2920221 = queryWeight, product of:
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.050563898 = queryNorm
              0.31904373 = fieldWeight in 180, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.0390625 = fieldNorm(doc=180)
          0.034253553 = weight(_text_:22 in 180) [ClassicSimilarity], result of:
            0.034253553 = score(doc=180,freq=2.0), product of:
              0.17706616 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050563898 = queryNorm
              0.19345059 = fieldWeight in 180, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=180)
      0.5 = coord(1/2)
    
    Abstract
    Purpose This investigation aims to examine the differences and similarities between activity tracking technology users from two regions (the USA and Germany) in their intended privacy-related behavior. The focus lies on data handling after hypothetical discontinuance of use, data protection and privacy policy seeking, and privacy concerns. Design/methodology/approach The data was collected through an online survey in 2019. In order to identify significant differences between participants from Germany and the USA, the chi-squared test and the Mann-Whitney U test were applied. Findings The intensity of several privacy-related concerns was significantly different between the two groups. The majority of the participants did not inform themselves about the respective data privacy policies or terms and conditions before installing an activity tracking application. The majority of the German participants knew that they could request the deletion of all their collected data. In contrast, only 35% out of 68 participants from the US knew about this option. Research limitations/implications This study intends to raise awareness about managing the collected health and fitness data after stopping to use activity tracking technologies. Furthermore, to reduce privacy and security concerns, the involvement of the government, companies and users is necessary to handle and share data more considerably and in a sustainable way. Originality/value This study sheds light on users of activity tracking technologies from a broad perspective (here, participants from the USA and Germany). It incorporates not only concerns and the privacy paradox but (intended) user behavior, including seeking information on data protection and privacy policy and handling data after hypothetical discontinuance of use of the technology.
    Date
    20. 1.2015 18:30:22
  12. Yu, C.; Xue, H.; An, L.; Li, G.: ¬A lightweight semantic-enhanced interactive network for efficient short-text matching (2023) 0.06
    0.06371069 = product of:
      0.12742138 = sum of:
        0.12742138 = sum of:
          0.09316782 = weight(_text_:light in 890) [ClassicSimilarity], result of:
            0.09316782 = score(doc=890,freq=2.0), product of:
              0.2920221 = queryWeight, product of:
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.050563898 = queryNorm
              0.31904373 = fieldWeight in 890, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.0390625 = fieldNorm(doc=890)
          0.034253553 = weight(_text_:22 in 890) [ClassicSimilarity], result of:
            0.034253553 = score(doc=890,freq=2.0), product of:
              0.17706616 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050563898 = queryNorm
              0.19345059 = fieldWeight in 890, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=890)
      0.5 = coord(1/2)
    
    Abstract
    Knowledge-enhanced short-text matching has been a significant task attracting much attention in recent years. However, the existing approaches cannot effectively balance effect and efficiency. Effective models usually consist of complex network structures leading to slow inference speed and the difficulties of applications in actual practice. In addition, most knowledge-enhanced models try to link the mentions in the text to the entities of the knowledge graphs-the difficulties of entity linking decrease the generalizability among different datasets. To address these problems, we propose a lightweight Semantic-Enhanced Interactive Network (SEIN) model for efficient short-text matching. Unlike most current research, SEIN employs an unsupervised method to select WordNet's most appropriate paraphrase description as the external semantic knowledge. It focuses on integrating semantic information and interactive information of text while simplifying the structure of other modules. We conduct intensive experiments on four real-world datasets, that is, Quora, Twitter-URL, SciTail, and SICK-E. Compared with state-of-the-art methods, SEIN achieves the best performance on most datasets. The experimental results proved that introducing external knowledge could effectively improve the performance of the short-text matching models. The research sheds light on the role of lightweight models in leveraging external knowledge to improve the effect of short-text matching.
    Date
    22. 1.2023 19:05:27
  13. Lavoie, B.; Connaway, L.S.; Dempsey, L.: Anatomy of aggregate collections : the example of Google print for libraries (2005) 0.06
    0.058687486 = product of:
      0.11737497 = sum of:
        0.11737497 = sum of:
          0.09682284 = weight(_text_:light in 1184) [ClassicSimilarity], result of:
            0.09682284 = score(doc=1184,freq=6.0), product of:
              0.2920221 = queryWeight, product of:
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.050563898 = queryNorm
              0.33156 = fieldWeight in 1184, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.0234375 = fieldNorm(doc=1184)
          0.02055213 = weight(_text_:22 in 1184) [ClassicSimilarity], result of:
            0.02055213 = score(doc=1184,freq=2.0), product of:
              0.17706616 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050563898 = queryNorm
              0.116070345 = fieldWeight in 1184, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0234375 = fieldNorm(doc=1184)
      0.5 = coord(1/2)
    
    Abstract
    Google's December 2004 announcement of its intention to collaborate with five major research libraries - Harvard University, the University of Michigan, Stanford University, the University of Oxford, and the New York Public Library - to digitize and surface their print book collections in the Google searching universe has, predictably, stirred conflicting opinion, with some viewing the project as a welcome opportunity to enhance the visibility of library collections in new environments, and others wary of Google's prospective role as gateway to these collections. The project has been vigorously debated on discussion lists and blogs, with the participating libraries commonly referred to as "the Google 5". One point most observers seem to concede is that the questions raised by this initiative are both timely and significant. The Google Print Library Project (GPLP) has galvanized a long overdue, multi-faceted discussion about library print book collections. The print book is core to library identity and practice, but in an era of zero-sum budgeting, it is almost inevitable that print book budgets will decline as budgets for serials, digital resources, and other materials expand. As libraries re-allocate resources to accommodate changing patterns of user needs, print book budgets may be adversely impacted. Of course, the degree of impact will depend on a library's perceived mission. A public library may expect books to justify their shelf-space, with de-accession the consequence of minimal use. A national library, on the other hand, has a responsibility to the scholarly and cultural record and may seek to collect comprehensively within particular areas, with the attendant obligation to secure the long-term retention of its print book collections. The combination of limited budgets, changing user needs, and differences in library collection strategies underscores the need to think about a collective, or system-wide, print book collection - in particular, how can an inter-institutional system be organized to achieve goals that would be difficult, and/or prohibitively expensive, for any one library to undertake individually [4]? Mass digitization programs like GPLP cast new light on these and other issues surrounding the future of library print book collections, but at this early stage, it is light that illuminates only dimly. It will be some time before GPLP's implications for libraries and library print book collections can be fully appreciated and evaluated. But the strong interest and lively debate generated by this initiative suggest that some preliminary analysis - premature though it may be - would be useful, if only to undertake a rough mapping of the terrain over which GPLP potentially will extend. At the least, some early perspective helps shape interesting questions for the future, when the boundaries of GPLP become settled, workflows for producing and managing the digitized materials become systematized, and usage patterns within the GPLP framework begin to emerge.
    This article offers some perspectives on GPLP in light of what is known about library print book collections in general, and those of the Google 5 in particular, from information in OCLC's WorldCat bibliographic database and holdings file. Questions addressed include: * Coverage: What proportion of the system-wide print book collection will GPLP potentially cover? What is the degree of holdings overlap across the print book collections of the five participating libraries? * Language: What is the distribution of languages associated with the print books held by the GPLP libraries? Which languages are predominant? * Copyright: What proportion of the GPLP libraries' print book holdings are out of copyright? * Works: How many distinct works are represented in the holdings of the GPLP libraries? How does a focus on works impact coverage and holdings overlap? * Convergence: What are the effects on coverage of using a different set of five libraries? What are the effects of adding the holdings of additional libraries to those of the GPLP libraries, and how do these effects vary by library type? These questions certainly do not exhaust the analytical possibilities presented by GPLP. More in-depth analysis might look at Google 5 coverage in particular subject areas; it also would be interesting to see how many books covered by the GPLP have already been digitized in other contexts. However, these questions are left to future studies. The purpose here is to explore a few basic questions raised by GPLP, and in doing so, provide an empirical context for the debate that is sure to continue for some time to come. A secondary objective is to lay some groundwork for a general set of questions that could be used to explore the implications of any mass digitization initiative. A suggested list of questions is provided in the conclusion of the article.
    Date
    26.12.2011 14:08:22
  14. ap: Mehr als 320 Millionen Seiten im Internet (1998) 0.06
    0.055900697 = product of:
      0.11180139 = sum of:
        0.11180139 = product of:
          0.22360279 = sum of:
            0.22360279 = weight(_text_:light in 833) [ClassicSimilarity], result of:
              0.22360279 = score(doc=833,freq=2.0), product of:
                0.2920221 = queryWeight, product of:
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.050563898 = queryNorm
                0.765705 = fieldWeight in 833, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.09375 = fieldNorm(doc=833)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Content
    Gegenüberstellung der Vollständigkeit verschiedener Suchmaschinen gemäß einer Untersuchung des Computerherstellers NEC: HotBot (34%); Altavista (28%); Northern Light (20%); Excite (14%), Lycos (3%)
  15. Krellenstein, M.: Document classification at Northern Light (1999) 0.06
    0.055900697 = product of:
      0.11180139 = sum of:
        0.11180139 = product of:
          0.22360279 = sum of:
            0.22360279 = weight(_text_:light in 4435) [ClassicSimilarity], result of:
              0.22360279 = score(doc=4435,freq=2.0), product of:
                0.2920221 = queryWeight, product of:
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.050563898 = queryNorm
                0.765705 = fieldWeight in 4435, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.09375 = fieldNorm(doc=4435)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  16. Schrodt, R.: Tiefen und Untiefen im wissenschaftlichen Sprachgebrauch (2008) 0.05
    0.053539254 = product of:
      0.10707851 = sum of:
        0.10707851 = product of:
          0.3212355 = sum of:
            0.3212355 = weight(_text_:3a in 140) [ClassicSimilarity], result of:
              0.3212355 = score(doc=140,freq=2.0), product of:
                0.42868128 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.050563898 = queryNorm
                0.7493574 = fieldWeight in 140, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0625 = fieldNorm(doc=140)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Content
    Vgl. auch: https://studylibde.com/doc/13053640/richard-schrodt. Vgl. auch: http%3A%2F%2Fwww.univie.ac.at%2FGermanistik%2Fschrodt%2Fvorlesung%2Fwissenschaftssprache.doc&usg=AOvVaw1lDLDR6NFf1W0-oC9mEUJf.
  17. Popper, K.R.: Three worlds : the Tanner lecture on human values. Deliverd at the University of Michigan, April 7, 1978 (1978) 0.05
    0.053539254 = product of:
      0.10707851 = sum of:
        0.10707851 = product of:
          0.3212355 = sum of:
            0.3212355 = weight(_text_:3a in 230) [ClassicSimilarity], result of:
              0.3212355 = score(doc=230,freq=2.0), product of:
                0.42868128 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.050563898 = queryNorm
                0.7493574 = fieldWeight in 230, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0625 = fieldNorm(doc=230)
          0.33333334 = coord(1/3)
      0.5 = coord(1/2)
    
    Source
    https%3A%2F%2Ftannerlectures.utah.edu%2F_documents%2Fa-to-z%2Fp%2Fpopper80.pdf&usg=AOvVaw3f4QRTEH-OEBmoYr2J_c7H
  18. Neumeier, F.: Verständnisprobleme : Internet Suchmaschinen (1998) 0.05
    0.052703682 = product of:
      0.105407365 = sum of:
        0.105407365 = product of:
          0.21081473 = sum of:
            0.21081473 = weight(_text_:light in 40) [ClassicSimilarity], result of:
              0.21081473 = score(doc=40,freq=4.0), product of:
                0.2920221 = queryWeight, product of:
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.050563898 = queryNorm
                0.7219136 = fieldWeight in 40, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.0625 = fieldNorm(doc=40)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Content
    Vorgestellt und bewertet werden: AltaVista (Note: 4); Excite (2); Hotbot (3); InfoSeek (3); Lycos (4); Northern Light (5); Open Text (5); WebCrawler (4); Yahoo (3)
    Object
    Northern Light
  19. Stock, M.; Stock, W.G.: Internet-Suchwerkzeuge im Vergleich (III) : Informationslinguistik und -statistik: AltaVista, FAST und Northern Light (2001) 0.05
    0.052703682 = product of:
      0.105407365 = sum of:
        0.105407365 = product of:
          0.21081473 = sum of:
            0.21081473 = weight(_text_:light in 5578) [ClassicSimilarity], result of:
              0.21081473 = score(doc=5578,freq=4.0), product of:
                0.2920221 = queryWeight, product of:
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.050563898 = queryNorm
                0.7219136 = fieldWeight in 5578, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.7753086 = idf(docFreq=372, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5578)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Suchmaschinen im World Wide Web arbeiten automatisch: Sie spüren Dokumente auf, indexieren sie, halten die Datenbank (mehr oder minder) aktuell und bieten den Kunden Retrievaloberflächen an. In unserem Known-Item-Retrievaltest (Password 11/2000) schnitten - in dieser Reihenfolge - Google, Alta Vista, Northern Light und FAST (All the Web) am besten ab. Die letzten drei Systeme arbeiten mit einer Kombination aus informationslinguistischen und informationsstatistischen Algorithmen, weshalb wir sie hier gemeinsam besprechen wollen. Im Zentrum unserer informationswissenschaftlichen Analysen stehen die "Highlights" der jeweiligen Suchwerkzeuge
  20. Hajibayova, L.; Jacob, E.K.: User-generated genre tags through the lens of genre theories (2014) 0.05
    0.05096855 = product of:
      0.1019371 = sum of:
        0.1019371 = sum of:
          0.07453426 = weight(_text_:light in 1450) [ClassicSimilarity], result of:
            0.07453426 = score(doc=1450,freq=2.0), product of:
              0.2920221 = queryWeight, product of:
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.050563898 = queryNorm
              0.255235 = fieldWeight in 1450, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.7753086 = idf(docFreq=372, maxDocs=44218)
                0.03125 = fieldNorm(doc=1450)
          0.027402842 = weight(_text_:22 in 1450) [ClassicSimilarity], result of:
            0.027402842 = score(doc=1450,freq=2.0), product of:
              0.17706616 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.050563898 = queryNorm
              0.15476047 = fieldWeight in 1450, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=1450)
      0.5 = coord(1/2)
    
    Abstract
    LIS genre studies have suggested that representing the genre of a resource could provide better knowledge representation, organization and retrieval (e.g., Andersen, 2008; Crowston & Kwasnik, 2003). Beghtol (2001) argues that genre analysis could be a useful tool for creating a "framework of analysis for a domain ... [to] structure and interpret texts, events, ideas, decisions, explanations and every other human activity in that domain" (p. 19). Although some studies of user-generated tagging vocabularies have found a preponderance of content-related tags (e.g., Munk & Mork, 2007), Lamere's (2008) study of the most frequently applied tags at Last.fm found that tags representing musical genres were favored by taggers. Studies of user-generated genre tags suggest that, unlike traditional indexing, which generally assigns a single genre, users' assignments of genre-related tags provide better representation of the fuzziness at the boundaries of genre categories (Inskip, 2009). In this way, user-generated genre tags are more in line with Bakhtin's (Bakhtin & Medvedev, 1928/1985) conceptualization of genre as an "aggregate of the means for seeing and conceptualizin reality" (p. 137). For Bakhtin (1986), genres are kinds of practice characterized by their "addressivity" (p. 95): Different genres correspond to different "conceptions of the addressee" and are "determined by that area of human activity and everyday life to which the given utterance is related" (p.95). Miller (1984) argues that genre refers to a "conventional category of discourse based in large-scale typification of rhetorical action; as action, it acquires meaning from situation and from the social context in which that situation arose" (p. 163). Genre is part of a social context that produces, reproduces, modifies and ultimately represents a particular text, but how to reunite genre and situation (or text and context) in systems of knowledge organization has not been addressed. Based on Devitt's (1993) argument suggesting that "our construction of genre is what helps us to construct a situation" (p. 577), one way to represent genre as "typified rhetorical actions based in recurrent situations" (Miller, 1984, p. 159) would be to employ genre tags generated by a particular group or community of users. This study suggests application of social network analysis to detect communities (Newman, 2006) of genre taggers and argues that communities of genre taggers can better define the nature and constitution of a discourse community while simultaneously shedding light on multifaceted representations of the resource genres.
    Source
    Knowledge organization in the 21st century: between historical patterns and future prospects. Proceedings of the Thirteenth International ISKO Conference 19-22 May 2014, Kraków, Poland. Ed.: Wieslaw Babik

Languages

Types

  • el 82
  • b 34
  • p 1
  • More… Less…

Themes