Search (68 results, page 1 of 4)

  • × language_ss:"e"
  • × theme_ss:"Internet"
  • × year_i:[2010 TO 2020}
  1. Egbert, J.; Biber, D.; Davies, M.: Developing a bottom-up, user-based method of web register classification (2015) 0.01
    0.00918054 = product of:
      0.04131243 = sum of:
        0.029360492 = product of:
          0.058720984 = sum of:
            0.058720984 = weight(_text_:web in 2158) [ClassicSimilarity], result of:
              0.058720984 = score(doc=2158,freq=16.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.6119082 = fieldWeight in 2158, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2158)
          0.5 = coord(1/2)
        0.011951938 = product of:
          0.023903877 = sum of:
            0.023903877 = weight(_text_:22 in 2158) [ClassicSimilarity], result of:
              0.023903877 = score(doc=2158,freq=2.0), product of:
                0.10297151 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02940506 = queryNorm
                0.23214069 = fieldWeight in 2158, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2158)
          0.5 = coord(1/2)
      0.22222222 = coord(2/9)
    
    Abstract
    This paper introduces a project to develop a reliable, cost-effective method for classifying Internet texts into register categories, and apply that approach to the analysis of a large corpus of web documents. To date, the project has proceeded in 2 key phases. First, we developed a bottom-up method for web register classification, asking end users of the web to utilize a decision-tree survey to code relevant situational characteristics of web documents, resulting in a bottom-up identification of register and subregister categories. We present details regarding the development and testing of this method through a series of 10 pilot studies. Then, in the second phase of our project we applied this procedure to a corpus of 53,000 web documents. An analysis of the results demonstrates the effectiveness of these methods for web register classification and provides a preliminary description of the types and distribution of registers on the web.
    Date
    4. 8.2015 19:22:04
  2. Joint, N.: Web 2.0 and the library : a transformational technology? (2010) 0.01
    0.0061203605 = product of:
      0.027541623 = sum of:
        0.019573662 = product of:
          0.039147325 = sum of:
            0.039147325 = weight(_text_:web in 4202) [ClassicSimilarity], result of:
              0.039147325 = score(doc=4202,freq=16.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.4079388 = fieldWeight in 4202, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4202)
          0.5 = coord(1/2)
        0.007967959 = product of:
          0.015935918 = sum of:
            0.015935918 = weight(_text_:22 in 4202) [ClassicSimilarity], result of:
              0.015935918 = score(doc=4202,freq=2.0), product of:
                0.10297151 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02940506 = queryNorm
                0.15476047 = fieldWeight in 4202, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4202)
          0.5 = coord(1/2)
      0.22222222 = coord(2/9)
    
    Abstract
    Purpose - This paper is the final one in a series which has tried to give an overview of so-called transformational areas of digital library technology. The aim has been to assess how much real transformation these applications can bring about, in terms of creating genuine user benefit and also changing everyday library practice. Design/methodology/approach - The paper provides a summary of some of the legal and ethical issues associated with web 2.0 applications in libraries, associated with a brief retrospective view of some relevant literature. Findings - Although web 2.0 innovations have had a massive impact on the larger World Wide Web, the practical impact on library service delivery has been limited to date. What probably can be termed transformational in the effect of web 2.0 developments on library and information work is their effect on some underlying principles of professional practice. Research limitations/implications - The legal and ethical challenges of incorporating web 2.0 platforms into mainstream institutional service delivery need to be subject to further research, so that the risks associated with these innovations are better understood at the strategic and policy-making level. Practical implications - This paper makes some recommendations about new principles of library and information practice which will help practitioners make better sense of these innovations in their overall information environment. Social implications - The paper puts in context some of the more problematic social impacts of web 2.0 innovations, without denying the undeniable positive contribution of social networking to the sphere of human interactivity. Originality/value - This paper raises some cautionary points about web 2.0 applications without adopting a precautionary approach of total prohibition. However, none of the suggestions or analysis in this piece should be considered to constitute legal advice. If such advice is required, the reader should consult appropriate legal professionals.
    Date
    22. 1.2011 17:54:04
  3. Oguz, F.; Koehler, W.: URL decay at year 20 : a research note (2016) 0.01
    0.005789892 = product of:
      0.026054513 = sum of:
        0.012110585 = product of:
          0.02422117 = sum of:
            0.02422117 = weight(_text_:web in 2651) [ClassicSimilarity], result of:
              0.02422117 = score(doc=2651,freq=2.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.25239927 = fieldWeight in 2651, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2651)
          0.5 = coord(1/2)
        0.013943928 = product of:
          0.027887857 = sum of:
            0.027887857 = weight(_text_:22 in 2651) [ClassicSimilarity], result of:
              0.027887857 = score(doc=2651,freq=2.0), product of:
                0.10297151 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02940506 = queryNorm
                0.2708308 = fieldWeight in 2651, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2651)
          0.5 = coord(1/2)
      0.22222222 = coord(2/9)
    
    Abstract
    All text is ephemeral. Some texts are more ephemeral than others. The web has proved to be among the most ephemeral and changing of information vehicles. The research note revisits Koehler's original data set after about 20 years since it was first collected. By late 2013, the number of URLs responding to a query had fallen to 1.6% of the original sample. A query of the 6 remaining URLs in February 2015 showed only 2 still responding.
    Date
    22. 1.2016 14:37:14
  4. Dalip, D.H.; Gonçalves, M.A.; Cristo, M.; Calado, P.: ¬A general multiview framework for assessing the quality of collaboratively created content on web 2.0 (2017) 0.00
    0.004931886 = product of:
      0.022193488 = sum of:
        0.012233539 = product of:
          0.024467077 = sum of:
            0.024467077 = weight(_text_:web in 3343) [ClassicSimilarity], result of:
              0.024467077 = score(doc=3343,freq=4.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.25496176 = fieldWeight in 3343, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3343)
          0.5 = coord(1/2)
        0.009959949 = product of:
          0.019919898 = sum of:
            0.019919898 = weight(_text_:22 in 3343) [ClassicSimilarity], result of:
              0.019919898 = score(doc=3343,freq=2.0), product of:
                0.10297151 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02940506 = queryNorm
                0.19345059 = fieldWeight in 3343, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3343)
          0.5 = coord(1/2)
      0.22222222 = coord(2/9)
    
    Date
    16.11.2017 13:04:22
    Object
    Web 2.0
  5. Stuart, D.: Web metrics for library and information professionals (2014) 0.00
    0.004308088 = product of:
      0.038772795 = sum of:
        0.038772795 = product of:
          0.07754559 = sum of:
            0.07754559 = weight(_text_:web in 2274) [ClassicSimilarity], result of:
              0.07754559 = score(doc=2274,freq=82.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.808072 = fieldWeight in 2274, product of:
                  9.055386 = tf(freq=82.0), with freq of:
                    82.0 = termFreq=82.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=2274)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    This is a practical guide to using web metrics to measure impact and demonstrate value. The web provides an opportunity to collect a host of different metrics, from those associated with social media accounts and websites to more traditional research outputs. This book is a clear guide for library and information professionals as to what web metrics are available and how to assess and use them to make informed decisions and demonstrate value. As individuals and organizations increasingly use the web in addition to traditional publishing avenues and formats, this book provides the tools to unlock web metrics and evaluate the impact of this content. The key topics covered include: bibliometrics, webometrics and web metrics; data collection tools; evaluating impact on the web; evaluating social media impact; investigating relationships between actors; exploring traditional publications in a new environment; web metrics and the web of data; the future of web metrics and the library and information professional. The book will provide a practical introduction to web metrics for a wide range of library and information professionals, from the bibliometrician wanting to demonstrate the wider impact of a researcher's work than can be demonstrated through traditional citations databases, to the reference librarian wanting to measure how successfully they are engaging with their users on Twitter. It will be a valuable tool for anyone who wants to not only understand the impact of content, but demonstrate this impact to others within the organization and beyond.
    Content
    1. Introduction. MetricsIndicators -- Web metrics and Ranganathan's laws of library science -- Web metrics for the library and information professional -- The aim of this book -- The structure of the rest of this book -- 2. Bibliometrics, webometrics and web metrics. Web metrics -- Information science metrics -- Web analytics -- Relational and evaluative metrics -- Evaluative web metrics -- Relational web metrics -- Validating the results -- 3. Data collection tools. The anatomy of a URL, web links and the structure of the web -- Search engines 1.0 -- Web crawlers -- Search engines 2.0 -- Post search engine 2.0: fragmentation -- 4. Evaluating impact on the web. Websites -- Blogs -- Wikis -- Internal metrics -- External metrics -- A systematic approach to content analysis -- 5. Evaluating social media impact. Aspects of social network sites -- Typology of social network sites -- Research and tools for specific sites and services -- Other social network sites -- URL shorteners: web analytic links on any site -- General social media impact -- Sentiment analysis -- 6. Investigating relationships between actors. Social network analysis methods -- Sources for relational network analysis -- 7. Exploring traditional publications in a new environment. More bibliographic items -- Full text analysis -- Greater context -- 8. Web metrics and the web of data. The web of data -- Building the semantic web -- Implications of the web of data for web metrics -- Investigating the web of data today -- SPARQL -- Sindice -- LDSpider: an RDF web crawler -- 9. The future of web metrics and the library and information professional. How far we have come -- The future of web metrics -- The future of the library and information professional and web metrics.
    RSWK
    Bibliothek / World Wide Web / World Wide Web 2.0 / Analyse / Statistik
    Bibliometrie / Semantic Web / Soziale Software
    Subject
    Bibliothek / World Wide Web / World Wide Web 2.0 / Analyse / Statistik
    Bibliometrie / Semantic Web / Soziale Software
  6. Yang, S.; Han, R.; Ding, J.; Song, Y.: ¬The distribution of Web citations (2012) 0.00
    0.0039954577 = product of:
      0.035959117 = sum of:
        0.035959117 = product of:
          0.071918234 = sum of:
            0.071918234 = weight(_text_:web in 2735) [ClassicSimilarity], result of:
              0.071918234 = score(doc=2735,freq=24.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.7494315 = fieldWeight in 2735, product of:
                  4.8989797 = tf(freq=24.0), with freq of:
                    24.0 = termFreq=24.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2735)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    A substantial amount of research has focused on the persistence or availability of Web citations. The present study analyzes Web citation distributions. Web citations are defined as the mentions of the URLs of Web pages (Web resources) as references in academic papers. The present paper primarily focuses on the analysis of the URLs of Web citations and uses three sets of data, namely, Set 1 from the Humanities and Social Science Index in China (CSSCI, 1998-2009), Set 2 from the publications of two international computer science societies, Communications of the ACM and IEEE Computer (1995-1999), and Set 3 from the medical science database, MEDLINE, of the National Library of Medicine (1994-2006). Web citation distributions are investigated based on Web site types, Web page types, URL frequencies, URL depths, URL lengths, and year of article publication. Results show significant differences in the Web citation distributions among the three data sets. However, when the URLs of Web citations with the same hostnames are aggregated, the distributions in the three data sets are consistent with the power law (the Lotka function).
  7. Spink, A.; Danby, S.; Mallan, K.; Butler, C.: Exploring young children's web searching and technoliteracy (2010) 0.00
    0.0033295476 = product of:
      0.029965928 = sum of:
        0.029965928 = product of:
          0.059931856 = sum of:
            0.059931856 = weight(_text_:web in 3623) [ClassicSimilarity], result of:
              0.059931856 = score(doc=3623,freq=24.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.6245262 = fieldWeight in 3623, product of:
                  4.8989797 = tf(freq=24.0), with freq of:
                    24.0 = termFreq=24.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3623)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    Purpose - This paper aims to report findings from an exploratory study investigating the web interactions and technoliteracy of children in the early childhood years. Previous research has studied aspects of older children's technoliteracy and web searching; however, few studies have analyzed web search data from children younger than six years of age. Design/methodology/approach - The study explored the Google web searching and technoliteracy of young children who are enrolled in a "preparatory classroom" or kindergarten (the year before young children begin compulsory schooling in Queensland, Australia). Young children were video- and audio-taped while conducting Google web searches in the classroom. The data were qualitatively analysed to understand the young children's web search behaviour. Findings - The findings show that young children engage in complex web searches, including keyword searching and browsing, query formulation and reformulation, relevance judgments, successive searches, information multitasking and collaborative behaviours. The study results provide significant initial insights into young children's web searching and technoliteracy. Practical implications - The use of web search engines by young children is an important research area with implications for educators and web technologies developers. Originality/value - This is the first study of young children's interaction with a web search engine.
  8. MacKay, B.; Watters, C.: ¬An examination of multisession web tasks (2012) 0.00
    0.0031877991 = product of:
      0.028690193 = sum of:
        0.028690193 = product of:
          0.057380386 = sum of:
            0.057380386 = weight(_text_:web in 255) [ClassicSimilarity], result of:
              0.057380386 = score(doc=255,freq=22.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.59793836 = fieldWeight in 255, product of:
                  4.690416 = tf(freq=22.0), with freq of:
                    22.0 = termFreq=22.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=255)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    Today, people perform many types of tasks on the web, including those that require multiple web sessions. In this article, we build on research about web tasks and present an in-depth evaluation of the types of tasks people perform on the web over multiple web sessions. Multisession web tasks are goal-based tasks that often contain subtasks requiring more than one web session to complete. We will detail the results of two longitudinal studies that we conducted to explore this topic. The first study was a weeklong web-diary study where participants self-reported information on their own multisession tasks. The second study was a monthlong field study where participants used a customized version of Firefox, which logged their interactions for both their own multisession tasks and their other web activity. The results from both studies found that people perform eight different types of multisession tasks, that these tasks often consist of several subtasks, that these lasted different lengths of time, and that users have unique strategies to help continue the tasks which involved a variety of web and browser tools such as search engines and bookmarks and external applications such as Notepad or Word. Using the results from these studies, we have suggested three guidelines for developers to consider when designing browser-tool features to help people perform these types of tasks: (a) to maintain a list of current multisession tasks, (b) to support multitasking, and (c) to manage task-related information between sessions.
  9. Oliveira Machado, L.M.; Souza, R.R.; Simões, M. da Graça: Semantic web or web of data? : a diachronic study (1999 to 2017) of the publications of Tim Berners-Lee and the World Wide Web Consortium (2019) 0.00
    0.0031877991 = product of:
      0.028690193 = sum of:
        0.028690193 = product of:
          0.057380386 = sum of:
            0.057380386 = weight(_text_:web in 5300) [ClassicSimilarity], result of:
              0.057380386 = score(doc=5300,freq=22.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.59793836 = fieldWeight in 5300, product of:
                  4.690416 = tf(freq=22.0), with freq of:
                    22.0 = termFreq=22.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5300)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    The web has been, in the last decades, the place where information retrieval achieved its maximum importance, given its ubiquity and the sheer volume of information. However, its exponential growth made the retrieval task increasingly hard, relying in its effectiveness on idiosyncratic and somewhat biased ranking algorithms. To deal with this problem, a "new" web, called the Semantic Web (SW), was proposed, bringing along concepts like "Web of Data" and "Linked Data," although the definitions and connections among these concepts are often unclear. Based on a qualitative approach built over a literature review, a definition of SW is presented, discussing the related concepts sometimes used as synonyms. It concludes that the SW is a comprehensive and ambitious construct that includes the great purpose of making the web a global database. It also follows the specifications developed and/or associated with its operationalization and the necessary procedures for the connection of data in an open format on the web. The goals of this comprehensive SW are the union of two outcomes still tenuously connected: the virtually unlimited possibility of connections between data-the web domain-with the potentiality of the automated inference of "intelligent" systems-the semantic component.
    Theme
    Semantic Web
  10. Bizer, C.; Mendes, P.N.; Jentzsch, A.: Topology of the Web of Data (2012) 0.00
    0.0031703631 = product of:
      0.028533269 = sum of:
        0.028533269 = product of:
          0.057066537 = sum of:
            0.057066537 = weight(_text_:web in 425) [ClassicSimilarity], result of:
              0.057066537 = score(doc=425,freq=34.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.59466785 = fieldWeight in 425, product of:
                  5.8309517 = tf(freq=34.0), with freq of:
                    34.0 = termFreq=34.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.03125 = fieldNorm(doc=425)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    The degree of structure of Web content is the determining factor for the types of functionality that search engines can provide. The more well structured the Web content is, the easier it is for search engines to understand Web content and provide advanced functionality, such as faceted filtering or the aggregation of content from multiple Web sites, based on this understanding. Today, most Web sites are generated from structured data that is stored in relational databases. Thus, it does not require too much extra effort for Web sites to publish this structured data directly on the Web in addition to HTML pages, and thus help search engines to understand Web content and provide improved functionality. An early approach to realize this idea and help search engines to understand Web content is Microformats, a technique for markingup structured data about specific types on entities-such as tags, blog posts, people, or reviews-within HTML pages. As Microformats are focused on a few entity types, the World Wide Web Consortium (W3C) started in 2004 to standardize RDFa as an alternative, more generic language for embedding any type of data into HTML pages. Today, major search engines such as Google, Yahoo, and Bing extract Microformat and RDFa data describing products, reviews, persons, events, and recipes from Web pages and use the extracted data to improve the user's search experience. The search engines have started to aggregate structured data from different Web sites and augment their search results with these aggregated information units in the form of rich snippets which combine, for instance, data This chapter gives an overview of the topology of the Web of Data that has been created by publishing data on the Web using the microformats RDFa, Microdata and Linked Data publishing techniques.
    Source
    Semantic search over the Web. Eds.: R. De Virgilio, et al
    Theme
    Semantic Web
  11. Derek Doran, D.; Gokhale, S.S.: ¬A classification framework for web robots (2012) 0.00
    0.0030757042 = product of:
      0.027681338 = sum of:
        0.027681338 = product of:
          0.055362675 = sum of:
            0.055362675 = weight(_text_:web in 505) [ClassicSimilarity], result of:
              0.055362675 = score(doc=505,freq=8.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.5769126 = fieldWeight in 505, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0625 = fieldNorm(doc=505)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    The behavior of modern web robots varies widely when they crawl for different purposes. In this article, we present a framework to classify these web robots from two orthogonal perspectives, namely, their functionality and the types of resources they consume. Applying the classification framework to a year-long access log from the UConn SoE web server, we present trends that point to significant differences in their crawling behavior.
  12. Fu, T.; Abbasi, A.; Chen, H.: ¬A focused crawler for Dark Web forums (2010) 0.00
    0.0028834727 = product of:
      0.025951253 = sum of:
        0.025951253 = product of:
          0.051902507 = sum of:
            0.051902507 = weight(_text_:web in 3471) [ClassicSimilarity], result of:
              0.051902507 = score(doc=3471,freq=18.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.5408555 = fieldWeight in 3471, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3471)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    The unprecedented growth of the Internet has given rise to the Dark Web, the problematic facet of the Web associated with cybercrime, hate, and extremism. Despite the need for tools to collect and analyze Dark Web forums, the covert nature of this part of the Internet makes traditional Web crawling techniques insufficient for capturing such content. In this study, we propose a novel crawling system designed to collect Dark Web forum content. The system uses a human-assisted accessibility approach to gain access to Dark Web forums. Several URL ordering features and techniques enable efficient extraction of forum postings. The system also includes an incremental crawler coupled with a recall-improvement mechanism intended to facilitate enhanced retrieval and updating of collected content. Experiments conducted to evaluate the effectiveness of the human-assisted accessibility approach and the recall-improvement-based, incremental-update procedure yielded favorable results. The human-assisted approach significantly improved access to Dark Web forums while the incremental crawler with recall improvement also outperformed standard periodic- and incremental-update approaches. Using the system, we were able to collect over 100 Dark Web forums from three regions. A case study encompassing link and content analysis of collected forums was used to illustrate the value and importance of gathering and analyzing content from such online communities.
  13. Mahesh, K.; Karanth, P.: ¬A novel knowledge organization scheme for the Web : superlinks with semantic roles (2012) 0.00
    0.0028834727 = product of:
      0.025951253 = sum of:
        0.025951253 = product of:
          0.051902507 = sum of:
            0.051902507 = weight(_text_:web in 822) [ClassicSimilarity], result of:
              0.051902507 = score(doc=822,freq=18.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.5408555 = fieldWeight in 822, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=822)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    We discuss the needs of a knowledge organization scheme for supporting Web-based software applications. We show how it differs from traditional knowledge organization schemes due to the virtual, dynamic, ad-hoc, userspecific and application-specific nature of Web-based knowledge. The sheer size of Web resources also adds to the complexity of organizing knowledge on the Web. As such, a standard, global scheme such as a single ontology for classifying and organizing all Web-based content is unrealistic. There is nevertheless a strong and immediate need for effective knowledge organization schemes to improve the efficiency and effectiveness of Web-based applications. In this context, we propose a novel knowledge organization scheme wherein concepts in the ontology of a domain are semantically interlinked with specific pieces of Web-based content using a rich hyper-linking structure known as Superlinks with well-defined semantic roles. We illustrate how such a knowledge organization scheme improves the efficiency and effectiveness of a Web-based e-commerce retail store.
  14. Maemura, E.; Worby, N.; Milligan, I.; Becker, C.: If these crawls could talk : studying and documenting web archives provenance (2018) 0.00
    0.0028834727 = product of:
      0.025951253 = sum of:
        0.025951253 = product of:
          0.051902507 = sum of:
            0.051902507 = weight(_text_:web in 4465) [ClassicSimilarity], result of:
              0.051902507 = score(doc=4465,freq=18.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.5408555 = fieldWeight in 4465, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4465)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    The increasing use and prominence of web archives raises the urgency of establishing mechanisms for transparency in the making of web archives to facilitate the process of evaluating a web archive's provenance, scoping, and absences. Some choices and process events are captured automatically, but their interactions are not currently well understood or documented. This study examined the decision space of web archives and its role in shaping what is and what is not captured in the web archiving process. By comparing how three different web archives collections were created and documented, we investigate how curatorial decisions interact with technical and external factors and we compare commonalities and differences. The findings reveal the need to understand both the social and technical context that shapes those decisions and the ways in which these individual decisions interact. Based on the study, we propose a framework for documenting key dimensions of a collection that addresses the situated nature of the organizational context, technical specificities, and unique characteristics of web materials that are the focus of a collection. The framework enables future researchers to undertake empirical work studying the process of creating web archives collections in different contexts.
  15. Rogers, R.: Digital methods (2013) 0.00
    0.0028770578 = product of:
      0.025893519 = sum of:
        0.025893519 = product of:
          0.051787037 = sum of:
            0.051787037 = weight(_text_:web in 2354) [ClassicSimilarity], result of:
              0.051787037 = score(doc=2354,freq=28.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.5396523 = fieldWeight in 2354, product of:
                  5.2915025 = tf(freq=28.0), with freq of:
                    28.0 = termFreq=28.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2354)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    In Digital Methods, Richard Rogers proposes a methodological outlook for social and cultural scholarly research on the Web that seeks to move Internet research beyond the study of online culture. It is not a toolkit for Internet research, or operating instructions for a software package; it deals with broader questions. How can we study social media to learn something about society rather than about social media use? How can hyperlinks reveal not just the value of a Web site but the politics of association? Rogers proposes repurposing Web-native techniques for research into cultural change and societal conditions. We can learn to reapply such "methods of the medium" as crawling and crowd sourcing, PageRank and similar algorithms, tag clouds and other visualizations; we can learn how they handle hits, likes, tags, date stamps, and other Web-native objects. By "thinking along" with devices and the objects they handle, digital research methods can follow the evolving methods of the medium. Rogers uses this new methodological outlook to examine the findings of inquiries into 9/11 search results, the recognition of climate change skeptics by climate-change-related Web sites, the events surrounding the Srebrenica massacre according to Dutch, Serbian, Bosnian, and Croatian Wikipedias, presidential candidates' social media "friends," and the censorship of the Iranian Web. With Digital Methods, Rogers introduces a new vision and method for Internet research and at the same time applies them to the Web's objects of study, from tiny particles (hyperlinks) to large masses (social media).
    Content
    The end of the virtual : digital methods -- The link and the politics of Web space -- The website as archived object -- Googlization and the inculpable engine -- Search as research -- National Web studies -- Social media and post-demographics -- Wikipedia as cultural reference -- After cyberspace : big data, small data.
    LCSH
    Web search engines
    World Wide Web / Research
    RSWK
    Internet / Recherche / World Wide Web 2.0
    Subject
    Internet / Recherche / World Wide Web 2.0
    Web search engines
    World Wide Web / Research
  16. Bhavnani, S.K.; Peck, F.A.: Scatter matters : regularities and implications for the scatter of healthcare information on the Web (2010) 0.00
    0.0028252148 = product of:
      0.025426934 = sum of:
        0.025426934 = product of:
          0.050853867 = sum of:
            0.050853867 = weight(_text_:web in 3433) [ClassicSimilarity], result of:
              0.050853867 = score(doc=3433,freq=12.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.5299281 = fieldWeight in 3433, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3433)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    Despite the development of huge healthcare Web sites and powerful search engines, many searchers end their searches prematurely with incomplete information. Recent studies suggest that users often retrieve incomplete information because of the complex scatter of relevant facts about a topic across Web pages. However, little is understood about regularities underlying such information scatter. To probe regularities within the scatter of facts across Web pages, this article presents the results of two analyses: (a) a cluster analysis of Web pages that reveals the existence of three page clusters that vary in information density and (b) a content analysis that suggests the role each of the above-mentioned page clusters play in providing comprehensive information. These results provide implications for the design of Web sites, search tools, and training to help users find comprehensive information about a topic and for a hypothesis describing the underlying mechanisms causing the scatter. We conclude by briefly discussing how the analysis of information scatter, at the granularity of facts, complements existing theories of information-seeking behavior.
  17. Gorgeon, A.; Swanson, E.B.: Web 2.0 according to Wikipedia : capturing an organizing vision (2011) 0.00
    0.0028252148 = product of:
      0.025426934 = sum of:
        0.025426934 = product of:
          0.050853867 = sum of:
            0.050853867 = weight(_text_:web in 4766) [ClassicSimilarity], result of:
              0.050853867 = score(doc=4766,freq=12.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.5299281 = fieldWeight in 4766, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4766)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    Is Web 2.0 more than a buzzword? In recent years, technologists and others have heatedly debated this question, even in Wikipedia, itself an example of Web 2.0. From the perspective of the present study, Web 2.0 may indeed be a buzzword, but more substantially it is also an example of an organizing vision that drives a community's discourse about certain new Information Technology (IT), serving to advance the technology's adoption and diffusion. Every organizing vision has a career that reflects its construction over time, and in the present study we examine Web 2.0's career as captured in its Wikipedia entry over a 5-year period, finding that it falls into three distinct periods termed Germination, Growth, and Maturation. The findings reveal how Wikipedia, as a discourse vehicle, treats new IT and its many buzzwords, and more broadly captures the careers of their organizing visions. Too, they further our understanding of Wikipedia as a new encyclopedic form, providing novel insights into its uses, its community of contributors, and their editing activities, as well as the dynamics of article construction.
    Object
    Web 2.0
  18. Spink, A.; Du, J.T.: Toward a Web search model : integrating multitasking, cognitive coordination, and cognitive shifts (2011) 0.00
    0.0027185641 = product of:
      0.024467077 = sum of:
        0.024467077 = product of:
          0.048934154 = sum of:
            0.048934154 = weight(_text_:web in 4624) [ClassicSimilarity], result of:
              0.048934154 = score(doc=4624,freq=16.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.5099235 = fieldWeight in 4624, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4624)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    Limited research has investigated the role of multitasking, cognitive coordination, and cognitive shifts during web search. Understanding these three behaviors is crucial to web search model development. This study aims to explore characteristics of multitasking behavior, types of cognitive shifts, and levels of cognitive coordination as well as the relationship between them during web search. Data collection included pre- and postquestionnaires, think-aloud protocols, web search logs, observations, and interviews with 42 graduate students who conducted 315 web search sessions with 221 information problems. Results show that web search is a dynamic interaction including the ordering of multiple information problems and the generation of evolving information problems, including task switching, multitasking, explicit task and implicit mental coordination, and cognitive shifting. Findings show that explicit task-level coordination is closely linked to multitasking, and implicit cognitive-level coordination is related to the task-coordination process; including information problem development and task switching. Coordination mechanisms directly result in cognitive state shifts including strategy, evaluation, and view states that affect users' holistic shifts in information problem understanding and knowledge contribution. A web search model integrating multitasking, cognitive coordination, and cognitive shifts (MCC model) is presented. Implications and further research also are discussed.
  19. Finnemann, N.: Web Archive (2019) 0.00
    0.0027185641 = product of:
      0.024467077 = sum of:
        0.024467077 = product of:
          0.048934154 = sum of:
            0.048934154 = weight(_text_:web in 5005) [ClassicSimilarity], result of:
              0.048934154 = score(doc=5005,freq=16.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.5099235 = fieldWeight in 5005, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5005)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    This article deals with the function of general web archives within the emerging organization of fast-growing digital knowledge resources. It opens with a brief overview of reasons why general web archives are needed. Sections two and three pre-sent major, long term web archive initiatives and discuss the purposes and possible functions and unknown future needs, demands and concerns. Section four analyses three main principles for the selection of materials to be preserved in contemporary web archiving strategies, topic-centric, domain-centric and time-centric archiving strategies and how to combine these to provide a broad and rich archive. Section five is concerned with inherent limitations and why web archives are always flawed. The last section deals with the question whether and how web archives may be considered a new type of knowledge organization system (KOS) necessary to preserve web materials, to allow for the development of a range of new methodologies, to analyze these particular corpora in long term and long tail perspectives, and to build a bridge towards the rapidly expanding but fragmented landscape of digital archives, libraries, research infrastructures and other sorts of digital repositories.
  20. Danowski, P.: Step one: blow up the silo! : Open bibliographic data, the first step towards Linked Open Data (2010) 0.00
    0.0025790567 = product of:
      0.023211509 = sum of:
        0.023211509 = product of:
          0.046423018 = sum of:
            0.046423018 = weight(_text_:web in 3962) [ClassicSimilarity], result of:
              0.046423018 = score(doc=3962,freq=10.0), product of:
                0.09596372 = queryWeight, product of:
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.02940506 = queryNorm
                0.48375595 = fieldWeight in 3962, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.2635105 = idf(docFreq=4597, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3962)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Abstract
    More and more libraries starting semantic web projects. The question about the license of the data is not discussed or the discussion is deferred to the end of project. In this paper is discussed why the question of the license is so important in context of the semantic web that is should be one of the first aspects in a semantic web project. Also it will be shown why a public domain weaver is the only solution that fulfill the the special requirements of the semantic web and that guaranties the reuseablitly of semantic library data for a sustainability of the projects.
    Object
    Web 2.0

Types

  • a 61
  • el 4
  • m 4
  • More… Less…