Search (639 results, page 32 of 32)

  • × theme_ss:"Internet"
  1. Broughton, V.; Lane, H.: Classification schemes revisited : applications to Web indexing and searching (2000) 0.00
    0.003298259 = product of:
      0.006596518 = sum of:
        0.006596518 = product of:
          0.046175625 = sum of:
            0.046175625 = weight(_text_:authors in 2476) [ClassicSimilarity], result of:
              0.046175625 = score(doc=2476,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.25184128 = fieldWeight in 2476, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2476)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Content
    A short discussion of using classification systems to organize the web, one of many such. The authors are both involved with BC2 and naturally think it is the best system for organizing information online. They list reasons why faceted classifications are best (e.g. no theoretical limits to specificity or exhaustivity; easier to handle complex subjects; flexible enough to accommodate different user needs) and take a brief look at how BC2 works. They conclude with a discussion of how and why it should be applied to online resources, and a plea for recognition of the importance of classification and subject analysis skills, even when full-text searching is available and databases respond instantly.
  2. Kuperman, V.: Productivity in the Internet mailing lists : a bibliometric analysis (2006) 0.00
    0.003298259 = product of:
      0.006596518 = sum of:
        0.006596518 = product of:
          0.046175625 = sum of:
            0.046175625 = weight(_text_:authors in 4907) [ClassicSimilarity], result of:
              0.046175625 = score(doc=4907,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.25184128 = fieldWeight in 4907, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4907)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Abstract
    The author examines patterns of productivity in the Internet mailing lists, also known as discussion lists or discussion groups. Datasets have been collected from electronic archives of two Internet mailing lists, the LINGUIST and the History of the English Language. Theoretical models widely used in informetric research have been applied to fit the distribution of posted messages over the population of authors. The Generalized Inverse Poisson-Gaussian and Poisson-lognormal distributions show excellent results in both datasets, while Lotka and Yule-Simon distribution demonstrate poor-to-mediocre fits. In the mailing list where moderation and quality control are enforced to a higher degree, i.e., the LINGUIST, Lotka, and Yule-Simon distributions perform better. The findings can be plausibly explained by the lesser applicability of the success-breedssuccess model to the information production in the electronic communication media, such as Internet mailing lists, where selectivity of publications is marginal or nonexistent. The hypothesis is preliminary, and needs to be validated against the larger variety of datasets. Characteristics of the quality control, competitiveness, and the reward structure in Internet mailing lists as compared to professional scholarly journals are discussed.
  3. Maharana, B.; Nayak, K.; Sahu, N.K.: Scholarly use of web resources in LIS research : a citation analysis (2006) 0.00
    0.003298259 = product of:
      0.006596518 = sum of:
        0.006596518 = product of:
          0.046175625 = sum of:
            0.046175625 = weight(_text_:authors in 53) [ClassicSimilarity], result of:
              0.046175625 = score(doc=53,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.25184128 = fieldWeight in 53, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=53)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Abstract
    Purpose - The essential purpose of this paper is to measure the amount of web resources used for scholarly contributions in the area of library and information science (LIS) in India. It further aims to make an analysis of the nature and type of web resources and studies the various standards for web citations. Design/methodology/approach - In this study, the result of analysis of 292 web citations spread over 95 scholarly papers published in the proceedings of the National Conference of the Society for Information Science, India (SIS-2005) has been reported. All the 292 web citations were scanned and data relating to types of web domains, file formats, styles of citations, etc., were collected through a structured check list. The data thus obtained were systematically analyzed, figurative representations were made and appropriate interpretations were drawn. Findings - The study revealed that 292 (34.88 per cent) out of 837 were web citations, proving a significant correlation between the use of Internet resources and research productivity of LIS professionals in India. The highest number of web citations (35.6 per cent) was from .edu/.ac type domains. Most of the web resources (46.9 per cent) cited in the study were hypertext markup language (HTML) files. Originality/value - The paper is the result of an original analysis of web citations undertaken in order to study the dependence of LIS professionals in India on web sources for their scholarly contributions. This carries research value for web content providers, authors and researchers in LIS.
  4. Goh, D.H.-L.; Ng, P.K.: Link decay in leading information science journals (2007) 0.00
    0.003298259 = product of:
      0.006596518 = sum of:
        0.006596518 = product of:
          0.046175625 = sum of:
            0.046175625 = weight(_text_:authors in 1334) [ClassicSimilarity], result of:
              0.046175625 = score(doc=1334,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.25184128 = fieldWeight in 1334, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1334)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Abstract
    Web citations have become common in scholarly publications as the amount of online literature increases. Yet, such links are not persistent and many decay over time, causing accessibility problems for readers. The present study investigates the link decay phenomenon in three leading information science journals. Articles spanning a period of 7 years (1997-2003) were downloaded, and their links were extracted. From these, a measure of link decay, the half-life, was computed to be approximately 5 years, which compares favorably against other disciplines (1.4-4.8 years). The study also investigated types of link accessibility errors encountered as well as examined characteristics of links that may be associated with decay. It was found that approximately 31% of all citations were not accessible during the time of testing, and the majority of errors were due to missing content (HTTP Error Code 404). Citations from the edu domain were also found to have the highest failure rates of 36% when compared with other popular top-level domains. Results indicate that link decay is a problem that cannot be ignored, and implications for journal authors and readers are discussed.
  5. Xiao, C.; Zhou, F.; Wu, Y.: Predicting audience gender in online content-sharing social networks (2013) 0.00
    0.003298259 = product of:
      0.006596518 = sum of:
        0.006596518 = product of:
          0.046175625 = sum of:
            0.046175625 = weight(_text_:authors in 954) [ClassicSimilarity], result of:
              0.046175625 = score(doc=954,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.25184128 = fieldWeight in 954, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=954)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Abstract
    Understanding the behavior and characteristics of web users is valuable when improving information dissemination, designing recommendation systems, and so on. In this work, we explore various methods of predicting the ratio of male viewers to female viewers on YouTube. First, we propose and examine two hypotheses relating to audience consistency and topic consistency. The former means that videos made by the same authors tend to have similar male-to-female audience ratios, whereas the latter means that videos with similar topics tend to have similar audience gender ratios. To predict the audience gender ratio before video publication, two features based on these two hypotheses and other features are used in multiple linear regression (MLR) and support vector regression (SVR). We find that these two features are the key indicators of audience gender, whereas other features, such as gender of the user and duration of the video, have limited relationships. Second, another method is explored to predict the audience gender ratio. Specifically, we use the early comments collected after video publication to predict the ratio via simple linear regression (SLR). The experiments indicate that this model can achieve better performance by using a few early comments. We also observe that the correlation between the number of early comments (cost) and the predictive accuracy (gain) follows the law of diminishing marginal utility. We build the functions of these elements via curve fitting to find the appropriate number of early comments (approximately 250) that can achieve maximum gain at minimum cost.
  6. Olson, N.; Nolin, J.M.; Nelhans, G.: Semantic web, ubiquitous computing, or internet of things? : a macro-analysis of scholarly publications (2015) 0.00
    0.003298259 = product of:
      0.006596518 = sum of:
        0.006596518 = product of:
          0.046175625 = sum of:
            0.046175625 = weight(_text_:authors in 2315) [ClassicSimilarity], result of:
              0.046175625 = score(doc=2315,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.25184128 = fieldWeight in 2315, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2315)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Abstract
    Purpose - The purpose of this paper is to investigate concepts that are used in depicting future visions of society, as afforded by technology, to map the extent of their use, examine the level of their dominance in different research areas and geographic boundaries, identify potential overlaps, analyse their longitudinal growth, and examine whether any of the identified concepts has assumed an overarching position. Design/methodology/approach - In total, 14 concepts, each of which is used to depict visions of future information infrastructures, were identified. More than 20,000 scholarly documents related to 11 of these concepts (those with 20 or more documents) are analysed by various qualitative/quantitative methods. Findings - The concepts most referred to are semantic web and ubiquitous computing (all years), and internet of things (Year 2013). Publications on some newer concepts (e.g. digital living, real world internet) are minimal. There are variations in the extent of use and preferred concepts based on geographic and disciplinary boundaries. The overlap in the use of these terms is minimal and none of these terms has assumed an overarching umbrella position. Research limitations/implications - This study is limited to scholarly publications; it would be relevant to also study the pattern of usage in governmental communications and policy documents. Social implications - By mapping multiplicity of concepts and the dispersion of discussions, the authors highlight the need for, and facilitate, a broader discussion of related social and societal implications. Originality/value - This paper is the first to present a collective of these related concepts and map the pattern of their occurrence and growth.
  7. Nori, R.: Web searching and navigation : age, intelligence, and familiarity (2020) 0.00
    0.003298259 = product of:
      0.006596518 = sum of:
        0.006596518 = product of:
          0.046175625 = sum of:
            0.046175625 = weight(_text_:authors in 5945) [ClassicSimilarity], result of:
              0.046175625 = score(doc=5945,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.25184128 = fieldWeight in 5945, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5945)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Abstract
    In using the Internet to solve everyday problems, older adults tend to find fewer correct answers compared to younger adults. Some authors have argued that these differences could be explained by age-related decline. The present study aimed to analyze the relationship between web-searching navigation and users' age, considering the Intelligence Quotient (IQ) and frequency of Internet and personal computer use. The intent was to identify differences due to age and not to other variables (that is, cognitive decline, expertise with the tool). Eighteen students (18-30?years) and 18 older adults (60-75?years) took part in the experiment. Inclusion criteria were the frequent use of computers and a web-searching activity; the older adults performed the Mini-Mental State Examination to exclude cognitive impairment. Participants were requested to perform the Kaufman Brief Intelligence Test 2nd ed. to measure their IQ level, and nine everyday web-searching tasks of differing complexity. The results showed that older participants spent more time on solving tasks than younger participants, but with the same accuracy as young people. Furthermore, nonverbal IQ improved performance in terms of time among the older participants. Age did not influence web-searching behavior in users with normal expertise and intelligence.
  8. XML data management : native XML and XML-enabled database systems (2003) 0.00
    0.0032316206 = product of:
      0.0064632413 = sum of:
        0.0064632413 = product of:
          0.045242686 = sum of:
            0.045242686 = weight(_text_:authors in 2073) [ClassicSimilarity], result of:
              0.045242686 = score(doc=2073,freq=12.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.24675304 = fieldWeight in 2073, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.015625 = fieldNorm(doc=2073)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Footnote
    There is some debate over what exactly constitutes a native XML database. Bourret (2003) favors the wider definition; other authors such as the Butler Group (2002) restrict the use of the term to databases systems designed and built solely for storage and manipulation of XML. Two examples of the lauer (Tamino and eXist) are covered in detailed chapters here but also included in this section is the embedded XML database system, Berkeley DB XML, considered by makers Sleepycat Software to be "native" in that it is capable of storing XML natively but built an top of the Berkeley DB engine. To the uninitiated, the revelation that schemas and DTDs are not required by either Tamino or eXist might seem a little strange. Tamino implements "loose coupling" where the validation behavior can be set to "strict," "lax" (i.e., apply only to parts of a document) or "skip" (no checking), in eXist, schemas are simply optional. Many DTDs and schemas evolve as the XML documents are acquired and so these may adhere to slightly different schemas, thus the database should support queries an similar documents that do not share the same structune. In fact, because of the difficulties in mappings between XML and database (especially relational) schemas native XML databases are very useful for storage of semi-structured data, a point not made in either chapter. The chapter an embedded databases represents a "third way," being neither native nor of the XML-enabled relational type. These databases run inside purpose-written applications and are accessed via an API or similar, meaning that the application developer does not need to access database files at the operating system level but can rely an supplied routines to, for example, fetch and update database records. Thus, end-users do not use the databases directly; the applications do not usually include ad hoc end-user query tools. This property renders embedded databases unsuitable for a large number of situations and they have become very much a niche market but this market is growing rapidly. Embedded databases share an address space with the application so the overhead of calls to the server is reduced, they also confer advantages in that they are easier to deploy, manage and administer compared to a conventional client-server solution. This chapter is a very good introduction to the subject, primers an generic embedded databases and embedded XML databases are helpfully provided before the author moves to an overview of the Open Source Berkeley system. Building an embedded database application makes far greater demands an the software developer and the remainder of the chapter is devoted to consideration of these programming issues.
    Relational database Management systems have been one of the great success stories of recent times and sensitive to the market, Most major vendors have responded by extending their products to handle XML data while still exploiting the range of facilities that a modern RDBMS affords. No book of this type would be complete without consideration of the "big these" (Oracle 9i, DB2, and SQL Server 2000 which each get a dedicated chapter) and though occasionally overtly piece-meal and descriptive the authors all note the shortcomings as well as the strengths of the respective systems. This part of the book is somewhat dichotomous, these chapters being followed by two that propose detailed solutions to somewhat theoretical problems, a generic architecture for storing XML in a RDBMS and using an object-relational approach to building an XML repository. The biography of the author of the latter (Paul Brown) contains the curious but strangely reassuring admission that "he remains puzzled by XML." The first five components are in-depth case studies of XMLdatabase applications. Necessarily diverse, few will be interested in all the topics presented but I was particularly interested in the first case study an bioinformatics. One of the twentieth century's greatest scientific undertakings was the Human Genome Project, the quest to list the information encoded by the sequence of DNA that makes up our genes and which has been referred to as "a paradigm for information Management in the life sciences" (Pearson & Soll, 1991). After a brief introduction to molecular biology to give the background to the information management problems, the authors turn to the use of XML in bioinformatics. Some of the data are hierarchical (e.g., the Linnaean classification of a human as a primate, primates as mammals, mammals are all vertebrates, etc.) but others are far more difficult to model. The Human Genome Project is virtually complete as far as the data acquisition phase is concerned and the immense volume of genome sequence data is no longer a very significant information Management issue per se. However bioinformaticians now need to interpret this information. Some data are relatively straightforward, e.g., the positioning of genes and sequence elements (e.g., promoters) within the sequences, but there is often little or no knowledge available an the direct and indirect interactions between them. There are vast numbers of such interrelationships; many complex data types and novel ones are constantly emerging, necessitating an extensible approach and the ability to manage semi-structured data. In the past, object databases such as AceDB (Durbin & Mieg, 1991) have gone some way to Meeting these aims but it is the combination of XML and databases that more completely addresses knowledge Management requirements of bioinformatics. XML is being enthusiastically adopted with a plethora of XML markup standards being developed, as authors Direen and Jones note "The unprecedented degree and flexibility of XML in terms of its ability to capture information is what makes it ideal for knowledge Management and for use in bioinformatics."
    After several detailed examples of XML, Direen and Jones discuss sequence comparisons. The ability to create scored comparisons by such techniques as sequence alignment is fundamental to bioinformatics. For example, the function of a gene product may be inferred from similarity with a gene of known function but originating from a different organism and any information modeling method must facilitate such comparisons. One such comparison tool, BLAST utilizes a heuristic method has become the tool of choice for many years and is integrated into the NeoCore XMS (XML Management System) described herein. Any set of sequences that can be identified using an XPath query may thus become the targets of an embedded search. Again examples are given, though a BLASTp (protein) search is labeled as being BLASTn (nucleotide sequence) in one of them. Some variants of BLAST are computationally intensive, e.g., tBLASTx where a nucleotide sequence is dynamically translated in all six reading frames and compared against similarly translated database sequences. Though these variants are implemented in NeoCore XMS, it would be interesting to see runtimes for such comparisons. Obviously the utility of this and the other four quite specific examples will depend an your interest in the application area but two that are more research-oriented and general follow them. These chapters (on using XML with inductive databases and an XML warehouses) are both readable critical reviews of their respective subject areas. For those involved in the implementation of performance-critical applications an examination of benchmark results is mandatory, however very few would examine the benchmark tests themselves. The picture that emerges from this section is that no single set is comprehensive and that some functionalities are not addressed by any available benchmark. As always, there is no Substitute for an intimate knowledge of your data and how it is used. In a direct comparison of an XML-enabled and a native XML database system (unfortunately neither is named), the authors conclude that though the native system has the edge in handling large documents this comes at the expense of increasing index and data file size. The need to use legacy data and software will certainly favor the all-pervasive XML-enabled RDBMS such as Oracle 9i and IBM's DB2. Of more general utility is the chapter by Schmauch and Fellhauer comparing the approaches used by database systems for the storing of XML documents. Many of the limitations of current XML-handling systems may be traced to problems caused by the semi-structured nature of the documents and while the authors have no panacea, the chapter forms a useful discussion of the issues and even raises the ugly prospect that a return to the drawing board may be unavoidable. The book concludes with an appraisal of the current status of XML by the editors that perhaps focuses a little too little an the database side but overall I believe this book to be very useful indeed. Some of the indexing is a little idiosyncratic, for example some tags used in the examples are indexed (perhaps a separate examples index would be better) and Ron Bourret's excellent web site might be better placed under "Bourret" rather than under "Ron" but this doesn't really detract from the book's qualities. The broad spectrum and careful balance of theory and practice is a combination that both database and XML professionals will find valuable."
  9. Johannsen, J.: InetBib 2004 in Bonn : Tagungsbericht: (2005) 0.00
    0.002724577 = product of:
      0.005449154 = sum of:
        0.005449154 = product of:
          0.010898308 = sum of:
            0.010898308 = weight(_text_:22 in 3125) [ClassicSimilarity], result of:
              0.010898308 = score(doc=3125,freq=2.0), product of:
                0.14084098 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.040219255 = queryNorm
                0.07738023 = fieldWeight in 3125, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.015625 = fieldNorm(doc=3125)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 1.2005 19:05:37
  10. Iyengar, S.S.: Visual based retrieval systems and Web mining (2001) 0.00
    0.0026386073 = product of:
      0.0052772146 = sum of:
        0.0052772146 = product of:
          0.0369405 = sum of:
            0.0369405 = weight(_text_:authors in 6520) [ClassicSimilarity], result of:
              0.0369405 = score(doc=6520,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.20147301 = fieldWeight in 6520, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.03125 = fieldNorm(doc=6520)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Abstract
    Relevance has been a difficult concept to define, let alone measure. In this paper, a simple operational definition of relevance is proposed for a Web-based library catalog: whether or not during a search session the user saves, prints, mails, or downloads a citation. If one of those actions is performed, the session is considered relevant to the user. An analysis is presented illustrating the advantages and disadvantages of this definition. With this definition and good transaction logging, it is possible to ascertain the relevance of a session. This was done for 905,970 sessions conducted with the University of California's Melvyl online catalog. Next, a methodology was developed to try to predict the relevance of a session. A number of variables were defined that characterize a session, none of which used any demographic information about the user. The values of the variables were computed for the sessions. Principal components analysis was used to extract a new set of variables out of the original set. A stratified random sampling technique was used to form ten strata such that each new strata of 90,570 sessions contained the same proportion of relevant to nonrelevant sessions. Logistic regression was used to ascertain the regression coefficients for nine of the ten strata. Then, the coefficients were used to predict the relevance of the sessions in the missing strata. Overall, 17.85% of the sessions were determined to be relevant. The predicted number of relevant sessions for all ten strata was 11 %, a 6.85% difference. The authors believe that the methodology can be further refined and the prediction improved. This methodology could also have significant application in improving user searching and also in predicting electronic commerce buying decisions without the use of personal demographic data
  11. Priss, U.: Alternatives to the "Semantic Web" : multi-strategy knowledge representation (2003) 0.00
    0.0026386073 = product of:
      0.0052772146 = sum of:
        0.0052772146 = product of:
          0.0369405 = sum of:
            0.0369405 = weight(_text_:authors in 2733) [ClassicSimilarity], result of:
              0.0369405 = score(doc=2733,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.20147301 = fieldWeight in 2733, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2733)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Abstract
    This paper argues that the Semantic Web needs to incorporate both formal and associative structures (and possibly a multitude of other structures and strategies) to be successful. The arguments for this claim are based on an observation of successes and failures in the areas of artificial intelligence (AI) and natural language processing (NLP). 1. Introduction The WWW provides numerous challenges for information and knowledge processing activities. Information may be available but not accessible or retrievable because of language barriers or insufficient search strategies. Data mining techniques may discover implicit information in explicit data but these techniques do not necessarily guarantee that the discovered information is relevant, significant and trustworthy. During the last several decades hundreds or thousands of computer and information scientists have developed probably thousands of natural language processing and artificial intelligence techniques that were aimed at solving problems related to intelligent information processing only to encounter more and more new obstacles along the way. The latest solution, the Semantic Web, appears as an open declaration of defeat: since natural language processing and AI techniques did not provide sufficient results, it is now proposed to put the burden an the shoulder of the authors of webpages who are expected to populate their pages with metadata and additional markup. Metadata is essentially a new form of controlled vocabulary; markup - at least in the form of XML, XSL, etc - is essentially a programming language. Existing studies of the use of controlled vocabularies and indexing practices in information science and studies of teaching programming languages to "everybody" (Python, 2002) have shown that both are difficult and full of unsolved problems. This can further dampen the expectations of the success of the Semantic Web. In contrast to machines and despite numerous inter-cultural conflicts around the world, humans do communicate surprisingly successfully even across national, linguistic and cultural boundaries. The question then arises: why are humans successful at information processing tasks such as information integration, translation and communication, which computers find so difficult? One obvious answer is that human cognition is embodied and grounded in our shared experiences of living in the same world. AI researchers have theoretically explored the idea of symbol grounding in the early 1990's but so far, connectionist artificial agents with perceptual interfaces have not been integrated with a large-scale capability of symbolic representations.
  12. Wouters, P.; Vries, R. de: Formally citing the Web (2004) 0.00
    0.0026386073 = product of:
      0.0052772146 = sum of:
        0.0052772146 = product of:
          0.0369405 = sum of:
            0.0369405 = weight(_text_:authors in 3093) [ClassicSimilarity], result of:
              0.0369405 = score(doc=3093,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.20147301 = fieldWeight in 3093, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3093)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Abstract
    How do authors refer to Web-based information sources in their formal scientific publications? It is not yet weIl known how scientists and scholars actually include new types of information sources, available through the new media, in their published work. This article reports an a comparative study of the lists of references in 38 scientific journals in five different scientific and social scientific fields. The fields are sociology, library and information science, biochemistry and biotechnology, neuroscience, and the mathematics of computing. As is weIl known, references, citations, and hyperlinks play different roles in academic publishing and communication. Our study focuses an hyperlinks as attributes of references in formal scholarly publications. The study developed and applied a method to analyze the differential roles of publishing media in the analysis of scientific and scholarly literature references. The present secondary databases that include reference and citation data (the Web of Science) cannot be used for this type of research. By the automated processing and analysis of the full text of scientific and scholarly articles, we were able to extract the references and hyperlinks contained in these references in relation to other features of the scientific and scholarly literature. Our findings show that hyperlinking references are indeed, as expected, abundantly present in the formal literature. They also tend to cite more recent literature than the average reference. The large majority of the references are to Web instances of traditional scientific journals. Other types of Web-based information sources are less weIl represented in the lists of references, except in the case of pure e-journals. We conclude that this can be explained by taking the role of the publisher into account. Indeed, it seems that the shift from print-based to electronic publishing has created new roles for the publisher. By shaping the way scientific references are hyperlinking to other information sources, the publisher may have a large impact an the availability of scientific and scholarly information.
  13. Stacey, Alison; Stacey, Adrian: Effective information retrieval from the Internet : an advanced user's guide (2004) 0.00
    0.0026386073 = product of:
      0.0052772146 = sum of:
        0.0052772146 = product of:
          0.0369405 = sum of:
            0.0369405 = weight(_text_:authors in 4497) [ClassicSimilarity], result of:
              0.0369405 = score(doc=4497,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.20147301 = fieldWeight in 4497, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4497)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Content
    Key Features - Importantly, the book enables readers to develop strategies which will continue to be useful despite the rapidly-evolving state of the Internet and Internet technologies - it is not about technological `tricks'. - Enables readers to be aware of and compensate for bias and errors which are ubiquitous an the Internet. - Provides contemporary information an the deficiencies in web skills of novice users as well as practical techniques for teaching such users. The Authors Dr Alison Stacey works at the Learning Resource Centre, Cambridge Regional College. Dr Adrian Stacey, formerly based at Cambridge University, is a software programmer. Readership The book is aimed at a wide range of librarians and other information professionals who need to retrieve information from the Internet efficiently, to evaluate their confidence in the information they retrieve and/or to train others to use the Internet. It is primarily aimed at intermediate to advanced users of the Internet. Contents Fundamentals of information retrieval from the Internet - why learn web searching technique; types of information requests; patterns for information retrieval; leveraging the technology: Search term choice: pinpointing information an the web - why choose queries carefully; making search terms work together; how to pick search terms; finding the 'unfindable': Blas an the Internet - importance of bias; sources of bias; usergenerated bias: selecting information with which you already agree; assessing and compensating for bias; case studies: Query reformulation and longer term strategies - how to interact with your search engine; foraging for information; long term information retrieval: using the Internet to find trends; automating searches: how to make your machine do your work: Assessing the quality of results- how to assess and ensure quality: The novice user and teaching internet skills - novice users and their problems with the web; case study: research in a college library; interpreting 'second hand' web information.
  14. Mansourian, Y.; Ford, N.: Web searchers' attributions of success and failure: an empirical study (2007) 0.00
    0.0026386073 = product of:
      0.0052772146 = sum of:
        0.0052772146 = product of:
          0.0369405 = sum of:
            0.0369405 = weight(_text_:authors in 840) [ClassicSimilarity], result of:
              0.0369405 = score(doc=840,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.20147301 = fieldWeight in 840, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.03125 = fieldNorm(doc=840)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Abstract
    Purpose - This paper reports the findings of a study designed to explore web searchers' perceptions of the causes of their search failure and success. In particular, it seeks to discover the extent to which the constructs locus of control and attribution theory might provide useful frameworks for understanding searchers' perceptions. Design/methodology/approach - A combination of inductive and deductive approaches were employed. Perceptions of failed and successful searches were derived from the inductive analysis of using open-ended qualitative interviews with a sample of 37 biologists at the University of Sheffield. These perceptions were classified into "internal" and "external" attributions, and the relationships between these categories and "successful" and "failed" searches were analysed deductively to test the extent to which they might be explainable using locus of control and attribution theory interpretive frameworks. Findings - All searchers were readily able to recall "successful" and "unsuccessful" searches. In a large majority of cases (82.4 per cent), they clearly attributed each search to either internal (e.g. ability or effort) or external (e.g. luck or information not being available) factors. The pattern of such relationships was analysed, and mapped onto those that would be predicted by locus of control and attribution theory. The authors conclude that the potential of these theoretical frameworks to illuminate one's understanding of web searching, and associated training, merits further systematic study. Research limitations/implications - The findings are based on a relatively small sample of academic and research staff in a particular subject area. Importantly, also, the study can at best provide a prima facie case for further systematic study since, although the patterns of attribution behaviour accord with those predictable by locus of control and attribution theory, data relating to the predictive elements of these theories (e.g. levels of confidence and achievement) were not available. This issue is discussed, and recommendations made for further work. Originality/value - The findings provide some empirical support for the notion that locus of control and attribution theory might - subject to the limitations noted above - be potentially useful theoretical frameworks for helping us better understand web-based information seeking. If so, they could have implications particularly for better understanding of searchers' motivations, and for the design and development of more effective search training programmes.
  15. Mansourian, Y.; Ford, N.: Search persistence and failure on the web : a "bounded rationality" and "satisficing" analysis (2007) 0.00
    0.0026386073 = product of:
      0.0052772146 = sum of:
        0.0052772146 = product of:
          0.0369405 = sum of:
            0.0369405 = weight(_text_:authors in 841) [ClassicSimilarity], result of:
              0.0369405 = score(doc=841,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.20147301 = fieldWeight in 841, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.03125 = fieldNorm(doc=841)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Abstract
    Purpose - This paper aims to examine our current knowledge of how searchers perceive and react to the possibility of missing potentially important information whilst searching the web is limited. The study reported here seeks to investigate such perceptions and reactions, and to explore the extent to which Simon's "bounded rationality" theory is useful in illuminating these issues. Design/methodology/approach - Totally 37 academic staff, research staff and research students in three university departments were interviewed about their web searching. The open-ended, semi-structured interviews were inductively analysed. Emergence of the concept of "good enough" searching prompted a further analysis to explore the extent to which the data could be interpreted in terms of Simon's concepts of "bounded rationality" and "satisficing". Findings - The results indicate that the risk of missing potentially important information was a matter of concern to the interviewees. Their estimations of the likely extent and importance of missed information affected decisions by individuals as to when to stop searching - decisions based on very different criteria, which map well onto Simon's concepts. On the basis of the interview data, the authors propose tentative categorizations of perceptions of the risk of missing information including "inconsequential" "tolerable" "damaging" and "disastrous" and search strategies including "perfunctory" "minimalist" "nervous" and "extensive". It is concluded that there is at least a prima facie case for bounded rationality and satisficing being considered as potentially useful concepts in our quest better to understand aspects of human information behaviour. Research limitations/implications - Although the findings are based on a relatively small sample and an exploratory qualitative analysis, it is argued that the study raises a number of interesting questions, and has implications for both the development of theory and practice in the areas of web searching and information literacy. Originality/value - The paper focuses on an aspect of web searching which has not to date been well explored. Whilst research has done much to illuminate searchers' perceptions of what they find on the web, we know relatively little of their perceptions of, and reactions to information that they fail to find. The study reported here provides some tentative models, based on empirical evidence, of these phenomena.
  16. Rosenfeld, L.; Morville, P.: Information architecture for the World Wide Web : designing large-scale Web sites (1998) 0.00
    0.0023087813 = product of:
      0.0046175625 = sum of:
        0.0046175625 = product of:
          0.032322936 = sum of:
            0.032322936 = weight(_text_:authors in 493) [ClassicSimilarity], result of:
              0.032322936 = score(doc=493,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.17628889 = fieldWeight in 493, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=493)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Abstract
    Some web sites "work" and some don't. Good web site consultants know that you can't just jump in and start writing HTML, the same way you can't build a house by just pouring a foundation and putting up some walls. You need to know who will be using the site, and what they'll be using it for. You need some idea of what you'd like to draw their attention to during their visit. Overall, you need a strong, cohesive vision for the site that makes it both distinctive and usable. Information Architecture for the World Wide Web is about applying the principles of architecture and library science to web site design. Each web site is like a public building, available for tourists and regulars alike to breeze through at their leisure. The job of the architect is to set up the framework for the site to make it comfortable and inviting for people to visit, relax in, and perhaps even return to someday. Most books on web development concentrate either on the aesthetics or the mechanics of the site. This book is about the framework that holds the two together. With this book, you learn how to design web sites and intranets that support growth, management, and ease of use. Special attention is given to: * The process behind architecting a large, complex site * Web site hierarchy design and organization Information Architecture for the World Wide Web is for webmasters, designers, and anyone else involved in building a web site. It's for novice web designers who, from the start, want to avoid the traps that result in poorly designed sites. It's for experienced web designers who have already created sites but realize that something "is missing" from their sites and want to improve them. It's for programmers and administrators who are comfortable with HTML, CGI, and Java but want to understand how to organize their web pages into a cohesive site. The authors are two of the principals of Argus Associates, a web consulting firm. At Argus, they have created information architectures for web sites and intranets of some of the largest companies in the United States, including Chrysler Corporation, Barron's, and Dow Chemical.
  17. Bruce, H.: ¬The user's view of the Internet (2002) 0.00
    0.0020434326 = product of:
      0.004086865 = sum of:
        0.004086865 = product of:
          0.00817373 = sum of:
            0.00817373 = weight(_text_:22 in 4344) [ClassicSimilarity], result of:
              0.00817373 = score(doc=4344,freq=2.0), product of:
                0.14084098 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.040219255 = queryNorm
                0.058035173 = fieldWeight in 4344, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.01171875 = fieldNorm(doc=4344)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Footnote
    Chapter 2 (Technology and People) focuses an several theories of technological acceptance and diffusion. Unfortunately, Bruce's presentation is somewhat confusing as he moves from one theory to next, never quite connecting them into a logical sequence or coherent whole. Two theories are of particular interest to Bruce: the Theory of Diffusion of Innovations and the Theory of Planned Behavior. The Theory of Diffusion of Innovations is an "information-centric view of technology acceptance" in which technology adopters are placed in the information flows of society from which they learn about innovations and "drive innovation adoption decisions" (p. 20). The Theory of Planned Behavior maintains that the "performance of a behavior is a joint function of intentions and perceived behavioral control" (i.e., how muck control a person thinks they have) (pp. 22-23). Bruce combines these two theories to form the basis for the Technology Acceptance Model. This model posits that "an individual's acceptance of information technology is based an beliefs, attitudes, intentions, and behaviors" (p. 24). In all these theories and models echoes a recurring theme: "individual perceptions of the innovation or technology are critical" in terms of both its characteristics and its use (pp. 24-25). From these, in turn, Bruce derives a predictive theory of the role personal perceptions play in technology adoption: Personal Innovativeness of Information Technology Adoption (PIITA). Personal inventiveness is defined as "the willingness of an individual to try out any new information technology" (p. 26). In general, the PIITA theory predicts that information technology will be adopted by individuals that have a greater exposure to mass media, rely less an the evaluation of information technology by others, exhibit a greater ability to cope with uncertainty and take risks, and requires a less positive perception of an information technology prior to its adoption. Chapter 3 (A Focus an Usings) introduces the User-Centered Paradigm (UCP). The UCP is characteristic of the shift of emphasis from technology to users as the driving force behind technology and research agendas for Internet development [for a dissenting view, see Andrew Dillion's (2003) challenge to the utility of user-centerness for design guidance]. It entails the "broad acceptance of the user-oriented perspective across a range of disciplines and professional fields," such as business, education, cognitive engineering, and information science (p. 34).
  18. Ronan, J.S.: Chat reference : A guide to live virtual reference services (2003) 0.00
    0.0019789555 = product of:
      0.003957911 = sum of:
        0.003957911 = product of:
          0.027705375 = sum of:
            0.027705375 = weight(_text_:authors in 2230) [ClassicSimilarity], result of:
              0.027705375 = score(doc=2230,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.15110476 = fieldWeight in 2230, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=2230)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Footnote
    Rez. in: JASIST 55(2004) no.4, S.366-368 (L. Mon): "The early years of the 21st century have been marked in the library community by a sharp increase in "chat reference" Services offered through library Web sites. Variously called "live reference," "virtual reference," or "chat reference," these Services utilize chat, MOO, instant messaging, and other online technologies to extend the live, synchronous interaction of the library reference desk to the Internet. In 2001, Stephen Francoeur found 272 libraries with chat reference Services worldwide, and in June 2003 Bernie Sloan's online listing of chat consortia included 47 consortia, which together represented more than 550 U.S. libraries. (Francoeur, 2001; Sloan, 2003) Jana Ronan's book is the latest entrant among recent works intended to guide librarians in the growing trend of setting up synchronous online "chat reference" services, and provides information about selecting software, setting service policies, conducting training, and managing staffing for a variety of different service configurations. Ronan, who serves as Interactive Reference Coordinator for the George A. Smathers Libraries at the University of Florida, Gainesville, covers the practical aspects of launching and managing chat reference services in 10 chapters, followed by five chapters of individual case studies contributed by other authors who describe chat reference implementations at SUNY Morrisville, Austin Peay State University, Paradise Valley Community College, University of South Florida, and the University of Florida.
  19. Net effects : how librarians can manage the unintended consequenees of the Internet (2003) 0.00
    0.0013193036 = product of:
      0.0026386073 = sum of:
        0.0026386073 = product of:
          0.01847025 = sum of:
            0.01847025 = weight(_text_:authors in 1796) [ClassicSimilarity], result of:
              0.01847025 = score(doc=1796,freq=2.0), product of:
                0.1833521 = queryWeight, product of:
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.040219255 = queryNorm
                0.10073651 = fieldWeight in 1796, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.558814 = idf(docFreq=1258, maxDocs=44218)
                  0.015625 = fieldNorm(doc=1796)
          0.14285715 = coord(1/7)
      0.5 = coord(1/2)
    
    Footnote
    Rez. in: JASIST 55(2004) no.11, S.1025-1026 (D.E. Agosto): ""Did you ever feel as though the Internet has caused you to lose control of your library?" So begins the introduction to this volume of over 50 articles, essays, library policies, and other documents from a variety of sources, most of which are library journals aimed at practitioners. Volume editor Block has a long history of library service as well as an active career as an online journalist. From 1977 to 1999 she was the Associate Director of Public Services at the St. Ambrose University library in Davenport, Iowa. She was also a Fox News Online weekly columnist from 1998 to 2000. She currently writes for and publishes the weekly ezine Exlibris, which focuses an the use of computers, the Internet, and digital databases to improve library services. Despite the promising premise of this book, the final product is largely a disappointment because of the superficial coverage of its issues. A listing of the most frequently represented sources serves to express the general level and style of the entries: nine articles are reprinted from Computers in Libraries, five from Library Journal, four from Library Journal NetConnect, four from ExLibris, four from American Libraries, three from College & Research Libraries News, two from Online, and two from The Chronicle of Higher Education. Most of the authors included contributed only one item, although Roy Tennant (manager of the California Digital Library) authored three of the pieces, and Janet L. Balas (library information systems specialist at the Monroeville Public Library in Pennsylvania) and Karen G. Schneider (coordinator of lii.org, the Librarians' Index to the Internet) each wrote two. Volume editor Block herself wrote six of the entries, most of which have been reprinted from ExLibris. Reading the volume is muck like reading an issue of one of these journals-a pleasant experience that discusses issues in the field without presenting much research. Net Effects doesn't offer much in the way of theory or research, but then again it doesn't claim to. Instead, it claims to be an "idea book" (p. 5) with practical solutions to Internet-generated library problems. While the idea is a good one, little of the material is revolutionary or surprising (or even very creative), and most of the solutions offered will already be familiar to most of the book's intended audience.

Years

Languages

  • d 316
  • e 309
  • f 9
  • el 1
  • m 1
  • ru 1
  • sp 1
  • More… Less…

Types

  • a 513
  • m 82
  • s 36
  • el 21
  • r 3
  • x 3
  • b 1
  • More… Less…

Subjects

Classifications