Search (44 results, page 1 of 3)

  • × theme_ss:"Internet"
  • × year_i:[2010 TO 2020}
  1. Egbert, J.; Biber, D.; Davies, M.: Developing a bottom-up, user-based method of web register classification (2015) 0.03
    0.033034686 = product of:
      0.15416187 = sum of:
        0.03496567 = weight(_text_:classification in 2158) [ClassicSimilarity], result of:
          0.03496567 = score(doc=2158,freq=6.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.3656675 = fieldWeight in 2158, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.046875 = fieldNorm(doc=2158)
        0.03496567 = weight(_text_:classification in 2158) [ClassicSimilarity], result of:
          0.03496567 = score(doc=2158,freq=6.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.3656675 = fieldWeight in 2158, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.046875 = fieldNorm(doc=2158)
        0.08423052 = sum of:
          0.059822515 = weight(_text_:texts in 2158) [ClassicSimilarity], result of:
            0.059822515 = score(doc=2158,freq=2.0), product of:
              0.16460659 = queryWeight, product of:
                5.4822793 = idf(docFreq=499, maxDocs=44218)
                0.03002521 = queryNorm
              0.36342722 = fieldWeight in 2158, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.4822793 = idf(docFreq=499, maxDocs=44218)
                0.046875 = fieldNorm(doc=2158)
          0.024408007 = weight(_text_:22 in 2158) [ClassicSimilarity], result of:
            0.024408007 = score(doc=2158,freq=2.0), product of:
              0.10514317 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03002521 = queryNorm
              0.23214069 = fieldWeight in 2158, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2158)
      0.21428572 = coord(3/14)
    
    Abstract
    This paper introduces a project to develop a reliable, cost-effective method for classifying Internet texts into register categories, and apply that approach to the analysis of a large corpus of web documents. To date, the project has proceeded in 2 key phases. First, we developed a bottom-up method for web register classification, asking end users of the web to utilize a decision-tree survey to code relevant situational characteristics of web documents, resulting in a bottom-up identification of register and subregister categories. We present details regarding the development and testing of this method through a series of 10 pilot studies. Then, in the second phase of our project we applied this procedure to a corpus of 53,000 web documents. An analysis of the results demonstrates the effectiveness of these methods for web register classification and provides a preliminary description of the types and distribution of registers on the web.
    Date
    4. 8.2015 19:22:04
  2. Johnson, E.H.: S R Ranganathan in the Internet age (2019) 0.03
    0.026648408 = product of:
      0.124359235 = sum of:
        0.02546139 = weight(_text_:subject in 5406) [ClassicSimilarity], result of:
          0.02546139 = score(doc=5406,freq=2.0), product of:
            0.10738805 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.03002521 = queryNorm
            0.23709705 = fieldWeight in 5406, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.046875 = fieldNorm(doc=5406)
        0.049448926 = weight(_text_:classification in 5406) [ClassicSimilarity], result of:
          0.049448926 = score(doc=5406,freq=12.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.5171319 = fieldWeight in 5406, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.046875 = fieldNorm(doc=5406)
        0.049448926 = weight(_text_:classification in 5406) [ClassicSimilarity], result of:
          0.049448926 = score(doc=5406,freq=12.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.5171319 = fieldWeight in 5406, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.046875 = fieldNorm(doc=5406)
      0.21428572 = coord(3/14)
    
    Abstract
    S R Ranganathan's ideas have influenced library classification since the inception of his Colon Classification in 1933. His address at Elsinore, "Library Classification Through a Century", was his grand vision of the century of progress in classification from 1876 to 1975, and looked to the future of faceted classification as the means to provide a cohesive system to organize the world's information. Fifty years later, the internet and its achievements, social ecology, and consequences present a far more complicated picture, with the library as he knew it as a very small part and the problems that he confronted now greatly exacerbated. The systematic nature of Ranganathan's canons, principles, postulates, and devices suggest that modern semantic algorithms could guide automatic subject tagging. The vision presented here is one of internet-wide faceted classification and retrieval, implemented as open, distributed facets providing unified faceted searching across all web sites.
  3. Danowski, P.: Step one: blow up the silo! : Open bibliographic data, the first step towards Linked Open Data (2010) 0.02
    0.01511595 = product of:
      0.0705411 = sum of:
        0.02018744 = weight(_text_:classification in 3962) [ClassicSimilarity], result of:
          0.02018744 = score(doc=3962,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.21111822 = fieldWeight in 3962, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.046875 = fieldNorm(doc=3962)
        0.030166224 = weight(_text_:bibliographic in 3962) [ClassicSimilarity], result of:
          0.030166224 = score(doc=3962,freq=2.0), product of:
            0.11688946 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.03002521 = queryNorm
            0.2580748 = fieldWeight in 3962, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.046875 = fieldNorm(doc=3962)
        0.02018744 = weight(_text_:classification in 3962) [ClassicSimilarity], result of:
          0.02018744 = score(doc=3962,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.21111822 = fieldWeight in 3962, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.046875 = fieldNorm(doc=3962)
      0.21428572 = coord(3/14)
    
    Content
    Vortrag im Rahmen der Session 93. Cataloguing der WORLD LIBRARY AND INFORMATION CONGRESS: 76TH IFLA GENERAL CONFERENCE AND ASSEMBLY, 10-15 August 2010, Gothenburg, Sweden - 149. Information Technology, Cataloguing, Classification and Indexing with Knowledge Management
  4. Bhatia, S.; Biyani, P.; Mitra, P.: Identifying the role of individual user messages in an online discussion and its use in thread retrieval (2016) 0.01
    0.012375482 = product of:
      0.05775225 = sum of:
        0.023791125 = weight(_text_:classification in 2650) [ClassicSimilarity], result of:
          0.023791125 = score(doc=2650,freq=4.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.24880521 = fieldWeight in 2650, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2650)
        0.023791125 = weight(_text_:classification in 2650) [ClassicSimilarity], result of:
          0.023791125 = score(doc=2650,freq=4.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.24880521 = fieldWeight in 2650, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2650)
        0.010170003 = product of:
          0.020340007 = sum of:
            0.020340007 = weight(_text_:22 in 2650) [ClassicSimilarity], result of:
              0.020340007 = score(doc=2650,freq=2.0), product of:
                0.10514317 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03002521 = queryNorm
                0.19345059 = fieldWeight in 2650, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2650)
          0.5 = coord(1/2)
      0.21428572 = coord(3/14)
    
    Abstract
    Online discussion forums have become a popular medium for users to discuss with and seek information from other users having similar interests. A typical discussion thread consists of a sequence of posts posted by multiple users. Each post in a thread serves a different purpose providing different types of information and, thus, may not be equally useful for all applications. Identifying the purpose and nature of each post in a discussion thread is thus an interesting research problem as it can help in improving information extraction and intelligent assistance techniques. We study the problem of classifying a given post as per its purpose in the discussion thread and employ features based on the post's content, structure of the thread, behavior of the participating users, and sentiment analysis of the post's content. We evaluate our approach on two forum data sets belonging to different genres and achieve strong classification performance. We also analyze the relative importance of different features used for the post classification task. Next, as a use case, we describe how the post class information can help in thread retrieval by incorporating this information in a state-of-the-art thread retrieval model.
    Date
    22. 1.2016 11:50:46
  5. Derek Doran, D.; Gokhale, S.S.: ¬A classification framework for web robots (2012) 0.01
    0.010875943 = product of:
      0.0761316 = sum of:
        0.0380658 = weight(_text_:classification in 505) [ClassicSimilarity], result of:
          0.0380658 = score(doc=505,freq=4.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.39808834 = fieldWeight in 505, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0625 = fieldNorm(doc=505)
        0.0380658 = weight(_text_:classification in 505) [ClassicSimilarity], result of:
          0.0380658 = score(doc=505,freq=4.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.39808834 = fieldWeight in 505, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0625 = fieldNorm(doc=505)
      0.14285715 = coord(2/14)
    
    Abstract
    The behavior of modern web robots varies widely when they crawl for different purposes. In this article, we present a framework to classify these web robots from two orthogonal perspectives, namely, their functionality and the types of resources they consume. Applying the classification framework to a year-long access log from the UConn SoE web server, we present trends that point to significant differences in their crawling behavior.
  6. Huang, C.; Fu, T.; Chen, H.: Text-based video content classification for online video-sharing sites (2010) 0.01
    0.009613066 = product of:
      0.06729146 = sum of:
        0.03364573 = weight(_text_:classification in 3452) [ClassicSimilarity], result of:
          0.03364573 = score(doc=3452,freq=8.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.35186368 = fieldWeight in 3452, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3452)
        0.03364573 = weight(_text_:classification in 3452) [ClassicSimilarity], result of:
          0.03364573 = score(doc=3452,freq=8.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.35186368 = fieldWeight in 3452, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3452)
      0.14285715 = coord(2/14)
    
    Abstract
    With the emergence of Web 2.0, sharing personal content, communicating ideas, and interacting with other online users in Web 2.0 communities have become daily routines for online users. User-generated data from Web 2.0 sites provide rich personal information (e.g., personal preferences and interests) and can be utilized to obtain insight about cyber communities and their social networks. Many studies have focused on leveraging user-generated information to analyze blogs and forums, but few studies have applied this approach to video-sharing Web sites. In this study, we propose a text-based framework for video content classification of online-video sharing Web sites. Different types of user-generated data (e.g., titles, descriptions, and comments) were used as proxies for online videos, and three types of text features (lexical, syntactic, and content-specific features) were extracted. Three feature-based classification techniques (C4.5, Naïve Bayes, and Support Vector Machine) were used to classify videos. To evaluate the proposed framework, user-generated data from candidate videos, which were identified by searching user-given keywords on YouTube, were first collected. Then, a subset of the collected data was randomly selected and manually tagged by users as our experiment data. The experimental results showed that the proposed approach was able to classify online videos based on users' interests with accuracy rates up to 87.2%, and all three types of text features contributed to discriminating videos. Support Vector Machine outperformed C4.5 and Naïve Bayes techniques in our experiments. In addition, our case study further demonstrated that accurate video-classification results are very useful for identifying implicit cyber communities on video-sharing Web sites.
  7. Oguz, F.; Koehler, W.: URL decay at year 20 : a research note (2016) 0.01
    0.0070192106 = product of:
      0.09826894 = sum of:
        0.09826894 = sum of:
          0.069792934 = weight(_text_:texts in 2651) [ClassicSimilarity], result of:
            0.069792934 = score(doc=2651,freq=2.0), product of:
              0.16460659 = queryWeight, product of:
                5.4822793 = idf(docFreq=499, maxDocs=44218)
                0.03002521 = queryNorm
              0.42399842 = fieldWeight in 2651, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                5.4822793 = idf(docFreq=499, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2651)
          0.02847601 = weight(_text_:22 in 2651) [ClassicSimilarity], result of:
            0.02847601 = score(doc=2651,freq=2.0), product of:
              0.10514317 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03002521 = queryNorm
              0.2708308 = fieldWeight in 2651, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2651)
      0.071428575 = coord(1/14)
    
    Abstract
    All text is ephemeral. Some texts are more ephemeral than others. The web has proved to be among the most ephemeral and changing of information vehicles. The research note revisits Koehler's original data set after about 20 years since it was first collected. By late 2013, the number of URLs responding to a query had fallen to 1.6% of the original sample. A query of the 6 remaining URLs in February 2015 showed only 2 still responding.
    Date
    22. 1.2016 14:37:14
  8. Gorrell, G.; Bontcheva, K.: Classifying Twitter favorites : Like, bookmark, or Thanks? (2016) 0.01
    0.0067974646 = product of:
      0.04758225 = sum of:
        0.023791125 = weight(_text_:classification in 2487) [ClassicSimilarity], result of:
          0.023791125 = score(doc=2487,freq=4.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.24880521 = fieldWeight in 2487, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2487)
        0.023791125 = weight(_text_:classification in 2487) [ClassicSimilarity], result of:
          0.023791125 = score(doc=2487,freq=4.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.24880521 = fieldWeight in 2487, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2487)
      0.14285715 = coord(2/14)
    
    Abstract
    Since its foundation in 2006, Twitter has enjoyed a meteoric rise in popularity, currently boasting over 500 million users. Its short text nature means that the service is open to a variety of different usage patterns, which have evolved rapidly in terms of user base and utilization. Prior work has categorized Twitter users, as well as studied the use of lists and re-tweets and how these can be used to infer user profiles and interests. The focus of this article is on studying why and how Twitter users mark tweets as "favorites"-a functionality with currently poorly understood usage, but strong relevance for personalization and information access applications. Firstly, manual analysis and classification are carried out on a randomly chosen set of favorited tweets, which reveal different approaches to using this functionality (i.e., bookmarks, thanks, like, conversational, and self-promotion). Secondly, an automatic favorites classification approach is proposed, based on the categories established in the previous step. Our machine learning experiments demonstrate a high degree of success in matching human judgments in classifying favorites according to usage type. In conclusion, we discuss the purposes to which these data could be put, in the context of identifying users' patterns of interests.
  9. Hannemann, J.; Kett, J.: Linked data for libraries (2010) 0.01
    0.00576784 = product of:
      0.04037488 = sum of:
        0.02018744 = weight(_text_:classification in 3964) [ClassicSimilarity], result of:
          0.02018744 = score(doc=3964,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.21111822 = fieldWeight in 3964, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.046875 = fieldNorm(doc=3964)
        0.02018744 = weight(_text_:classification in 3964) [ClassicSimilarity], result of:
          0.02018744 = score(doc=3964,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.21111822 = fieldWeight in 3964, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.046875 = fieldNorm(doc=3964)
      0.14285715 = coord(2/14)
    
    Content
    Vortrag im Rahmen der Session 93. Cataloguing der WORLD LIBRARY AND INFORMATION CONGRESS: 76TH IFLA GENERAL CONFERENCE AND ASSEMBLY, 10-15 August 2010, Gothenburg, Sweden - 149. Information Technology, Cataloguing, Classification and Indexing with Knowledge Management
  10. Zubiaga, A.; Spina, D.; Martínez, R.; Fresno, V.: Real-time classification of Twitter trends (2015) 0.01
    0.00576784 = product of:
      0.04037488 = sum of:
        0.02018744 = weight(_text_:classification in 1661) [ClassicSimilarity], result of:
          0.02018744 = score(doc=1661,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.21111822 = fieldWeight in 1661, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.046875 = fieldNorm(doc=1661)
        0.02018744 = weight(_text_:classification in 1661) [ClassicSimilarity], result of:
          0.02018744 = score(doc=1661,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.21111822 = fieldWeight in 1661, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.046875 = fieldNorm(doc=1661)
      0.14285715 = coord(2/14)
    
  11. Schmidt, E.; Cohen, J.: ¬Die Vernetzung der Welt : ein Blick in unsere Zukunft (2013) 0.01
    0.0051063383 = product of:
      0.07148873 = sum of:
        0.07148873 = weight(_text_:henry in 3338) [ClassicSimilarity], result of:
          0.07148873 = score(doc=3338,freq=2.0), product of:
            0.23560001 = queryWeight, product of:
              7.84674 = idf(docFreq=46, maxDocs=44218)
              0.03002521 = queryNorm
            0.30343264 = fieldWeight in 3338, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.84674 = idf(docFreq=46, maxDocs=44218)
              0.02734375 = fieldNorm(doc=3338)
      0.071428575 = coord(1/14)
    
    Footnote
    Pressestimmen - In diesem faszinierenden Buch machen Eric Schmidt und Jared Cohen von ihrer einzigartigen Sachkenntnis Gebrauch, um uns eine Zukunft auszumalen, in der die Einkommen steigen, die Partizipation zunimmt und ein echter Sinn für Gemeinschaft entsteht - vorausgesetzt, wir treffen heute die richtigen Entscheidungen. (Bill Clinton) - Dieses Buch erklärt sowohl, was die neue Welt ausmacht, die das Internet schafft, als auch die Herausforderungen, die sie mit sich bringt. Niemand könnte das besser als Eric Schmidt und Jared Cohen. (Tony Blair) - Selbst wer nicht alle Schlussfolgerungen teilen mag, wird viel von diesem anregenden Buch lernen. (Henry A. Kissinger) - Auf dieses Buch habe ich gewartet: Eine prägnante und überzeugende Darstellung der Auswirkungen, die Technologie auf Krieg und Frieden, Freiheit und Diplomatie hat ... - Eine unverzichtbare Lektüre. (Madeleine Albright) -Dies ist das wichtigste - und faszinierendste - Buch, das bislang über die Auswirkungen des Digitalzeitalters auf unsere Welt geschrieben wurde. (Walter Isaacson) - «Die Vernetzung der Welt» verbindet auf faszinierende Weise Konzepte und Einblicke darüber, wie die sich die virtuelle Welt und die internationale Staatenordnung durchkreuzen. (Robert B. Zoellick) - Kaum jemand auf der Welt beschäftigt sich mehr damit, sich das neue Digitalzeitalter auszumalen - und es zu gestalten - als Eric Schmidt und Jared Cohen. Mit diesem Buch werfen sie einen Blick in ihre Kristallkugel und laden uns ein, ihnen dabei über die Schulter zu schauen. (Michael Bloomberg) - Dieses Buch ist die aufschlussreichste Erkundung unserer Zukunft, die ich je gelesen habe. Ich konnte es gar nicht mehr weglegen. (Sir Richard Branson) - «Die Vernetzung der Welt» ist Pflichtlektüre für alle, die das Ausmaß der digitalen Revolution wirklich verstehen wollen. (General Michael Hayden - ehemaliger Direktor der CIA) - Trotz der Herkunft der Autoren verbreitet «Die Vernetzung der Welt» keine Silicon-Valley-Propaganda ... Und was noch wichtiger ist: Es hebt die Debatte über Technologie auf ein höheres Niveau - weg vom banalen Streit über den Nutzen von Dating-Apps, hin zu allgemeineren Frage nach der gegenseitigen Beeinflussung von Technologie und Macht. (The Economist) - Dieses Buch ist deutlich mehr als nur Science Fiction. Es diskutiert hellsichtig und offen die entscheidenden Fragen, denen wir uns schon jetzt stellen müssen. Wer die Welt der Zukunft verstehen will, sollte es daher unbedingt lesen. (NDR Kultur)
  12. Segev, E.: Google and the digital divide : the bias of online knowledge (2010) 0.00
    0.004806533 = product of:
      0.03364573 = sum of:
        0.016822865 = weight(_text_:classification in 3079) [ClassicSimilarity], result of:
          0.016822865 = score(doc=3079,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.17593184 = fieldWeight in 3079, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3079)
        0.016822865 = weight(_text_:classification in 3079) [ClassicSimilarity], result of:
          0.016822865 = score(doc=3079,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.17593184 = fieldWeight in 3079, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3079)
      0.14285715 = coord(2/14)
    
    Abstract
    Aimed at information and communication professionals, scholars and students, Google and the Digital Divide: The Biases of Online Knowledge provides invaluable insight into the significant role that search engines play in growing the digital divide between individuals, organizations, and states. With a specific focus on Google, author Elad Segev explains the concept of the digital divide and the effects that today's online environment has on knowledge bias, power, and control. Using innovative methods and research approaches, Segev compares the popular search queries in Google and Yahoo in the United States and other countries and analyzes the various biases in Google News and Google Earth. Google and the Digital Divide shows the many ways in which users manipulate Google's information across different countries, as well as dataset and classification systems, economic and political value indexes, specific search indexes, locality of use indexes, and much more. Segev presents important new social and political perspectives to illustrate the challenges brought about by search engines, and explains the resultant political, communicative, commercial, and international implications.
  13. Yang, M.; Kiang, M.; Chen, H.; Li, Y.: Artificial immune system for illicit content identification in social media (2012) 0.00
    0.004806533 = product of:
      0.03364573 = sum of:
        0.016822865 = weight(_text_:classification in 4980) [ClassicSimilarity], result of:
          0.016822865 = score(doc=4980,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.17593184 = fieldWeight in 4980, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4980)
        0.016822865 = weight(_text_:classification in 4980) [ClassicSimilarity], result of:
          0.016822865 = score(doc=4980,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.17593184 = fieldWeight in 4980, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4980)
      0.14285715 = coord(2/14)
    
    Abstract
    Social media is frequently used as a platform for the exchange of information and opinions as well as propaganda dissemination. But online content can be misused for the distribution of illicit information, such as violent postings in web forums. Illicit content is highly distributed in social media, while non-illicit content is unspecific and topically diverse. It is costly and time consuming to label a large amount of illicit content (positive examples) and non-illicit content (negative examples) to train classification systems. Nevertheless, it is relatively easy to obtain large volumes of unlabeled content in social media. In this article, an artificial immune system-based technique is presented to address the difficulties in the illicit content identification in social media. Inspired by the positive selection principle in the immune system, we designed a novel labeling heuristic based on partially supervised learning to extract high-quality positive and negative examples from unlabeled datasets. The empirical evaluation results from two large hate group web forums suggest that our proposed approach generally outperforms the benchmark techniques and exhibits more stable performance.
  14. Kalman, Y.M.; Ravid, G.: Filing, piling, and everything in between : the dynamics of E-mail inbox management (2015) 0.00
    0.004806533 = product of:
      0.03364573 = sum of:
        0.016822865 = weight(_text_:classification in 2336) [ClassicSimilarity], result of:
          0.016822865 = score(doc=2336,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.17593184 = fieldWeight in 2336, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2336)
        0.016822865 = weight(_text_:classification in 2336) [ClassicSimilarity], result of:
          0.016822865 = score(doc=2336,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.17593184 = fieldWeight in 2336, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=2336)
      0.14285715 = coord(2/14)
    
    Abstract
    Managing the constant flow of incoming messages is a daily challenge faced by knowledge workers who use technologies such as e-mail and other digital communication tools. This study focuses on the most ubiquitous of these technologies, e-mail, and unobtrusively explores the ongoing inbox-management activities of thousands of users worldwide over a period of 8 months. The study describes the dynamics of these inboxes throughout the day and the week as users strive to handle incoming messages, read them, classify them, respond to them in a timely manner, and archive them for future reference, all while carrying out the daily tasks of knowledge workers. It then tests several hypotheses about the influence of specific inbox-management behaviors in mitigating the causes of e-mail overload, and proposes a continuous index that quantifies one of these inbox-management behaviors. This inbox clearing index (ICI) expands on the widely cited trichotomous classification of users into frequent filers, spring cleaners, and no filers, as suggested by Whittaker and Sidner (1996). We propose that the ICI allows shifting the focus, from classifying users to characterizing a diversity of user behaviors and measuring the relationships between these behaviors and desired outcomes.
  15. Thelwall, M.; Goriunova, O.; Vis, F.; Faulkner, S.; Burns, A.; Aulich, J.; Mas-Bleda, A.; Stuart, E.; D'Orazio, F.: Chatting through pictures : a classification of images tweeted in one week in the UK and USA (2016) 0.00
    0.004806533 = product of:
      0.03364573 = sum of:
        0.016822865 = weight(_text_:classification in 3215) [ClassicSimilarity], result of:
          0.016822865 = score(doc=3215,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.17593184 = fieldWeight in 3215, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3215)
        0.016822865 = weight(_text_:classification in 3215) [ClassicSimilarity], result of:
          0.016822865 = score(doc=3215,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.17593184 = fieldWeight in 3215, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3215)
      0.14285715 = coord(2/14)
    
  16. Zielinski, K.; Nielek, R.; Wierzbicki, A.; Jatowt, A.: Computing controversy : formal model and algorithms for detecting controversy on Wikipedia and in search queries (2018) 0.00
    0.004806533 = product of:
      0.03364573 = sum of:
        0.016822865 = weight(_text_:classification in 5093) [ClassicSimilarity], result of:
          0.016822865 = score(doc=5093,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.17593184 = fieldWeight in 5093, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5093)
        0.016822865 = weight(_text_:classification in 5093) [ClassicSimilarity], result of:
          0.016822865 = score(doc=5093,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.17593184 = fieldWeight in 5093, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5093)
      0.14285715 = coord(2/14)
    
    Abstract
    Controversy is a complex concept that has been attracting attention of scholars from diverse fields. In the era of Internet and social media, detecting controversy and controversial concepts by the means of automatic methods is especially important. Web searchers could be alerted when the contents they consume are controversial or when they attempt to acquire information on disputed topics. Presenting users with the indications and explanations of the controversy should offer them chance to see the "wider picture" rather than letting them obtain one-sided views. In this work we first introduce a formal model of controversy as the basis of computational approaches to detecting controversial concepts. Then we propose a classification based method for automatic detection of controversial articles and categories in Wikipedia. Next, we demonstrate how to use the obtained results for the estimation of the controversy level of search queries. The proposed method can be incorporated into search engines as a component responsible for detection of queries related to controversial topics. The method is independent of the search engine's retrieval and search results recommendation algorithms, and is therefore unaffected by a possible filter bubble. Our approach can be also applied in Wikipedia or other knowledge bases for supporting the detection of controversy and content maintenance. Finally, we believe that our results could be useful for social science researchers for understanding the complex nature of controversy and in fostering their studies.
  17. Rodríguez-Vidal, J.; Gonzalo, J.; Plaza, L.; Anaya Sánchez, H.: Automatic detection of influencers in social networks : authority versus domain signals (2019) 0.00
    0.004806533 = product of:
      0.03364573 = sum of:
        0.016822865 = weight(_text_:classification in 5301) [ClassicSimilarity], result of:
          0.016822865 = score(doc=5301,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.17593184 = fieldWeight in 5301, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5301)
        0.016822865 = weight(_text_:classification in 5301) [ClassicSimilarity], result of:
          0.016822865 = score(doc=5301,freq=2.0), product of:
            0.09562149 = queryWeight, product of:
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.03002521 = queryNorm
            0.17593184 = fieldWeight in 5301, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.1847067 = idf(docFreq=4974, maxDocs=44218)
              0.0390625 = fieldNorm(doc=5301)
      0.14285715 = coord(2/14)
    
    Abstract
    Given the task of finding influencers (opinion makers) for a given domain in a social network, we investigate (a) what is the relative importance of domain and authority signals, (b) what is the most effective way of combining signals (voting, classification, learning to rank, etc.) and how best to model the vocabulary signal, and (c) how large is the gap between supervised and unsupervised methods and what are the practical consequences. Our best results on the RepLab dataset (which improves the state of the art) uses language models to learn the domain-specific vocabulary used by influencers and combines domain and authority models using a Learning to Rank algorithm. Our experiments show that (a) both authority and domain evidence can be trained from the vocabulary of influencers; (b) once the language of influencers is modeled as a likelihood signal, further supervised learning and additional network-based signals only provide marginal improvements; and (c) the availability of training data sets is crucial to obtain competitive results in the task. Our most remarkable finding is that influencers do use a distinctive vocabulary, which is a more reliable signal than nontextual network indicators such as the number of followers, retweets, and so on.
  18. Joint, N.: Web 2.0 and the library : a transformational technology? (2010) 0.00
    0.0035871807 = product of:
      0.025110263 = sum of:
        0.016974261 = weight(_text_:subject in 4202) [ClassicSimilarity], result of:
          0.016974261 = score(doc=4202,freq=2.0), product of:
            0.10738805 = queryWeight, product of:
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.03002521 = queryNorm
            0.15806471 = fieldWeight in 4202, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.576596 = idf(docFreq=3361, maxDocs=44218)
              0.03125 = fieldNorm(doc=4202)
        0.008136002 = product of:
          0.016272005 = sum of:
            0.016272005 = weight(_text_:22 in 4202) [ClassicSimilarity], result of:
              0.016272005 = score(doc=4202,freq=2.0), product of:
                0.10514317 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03002521 = queryNorm
                0.15476047 = fieldWeight in 4202, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=4202)
          0.5 = coord(1/2)
      0.14285715 = coord(2/14)
    
    Abstract
    Purpose - This paper is the final one in a series which has tried to give an overview of so-called transformational areas of digital library technology. The aim has been to assess how much real transformation these applications can bring about, in terms of creating genuine user benefit and also changing everyday library practice. Design/methodology/approach - The paper provides a summary of some of the legal and ethical issues associated with web 2.0 applications in libraries, associated with a brief retrospective view of some relevant literature. Findings - Although web 2.0 innovations have had a massive impact on the larger World Wide Web, the practical impact on library service delivery has been limited to date. What probably can be termed transformational in the effect of web 2.0 developments on library and information work is their effect on some underlying principles of professional practice. Research limitations/implications - The legal and ethical challenges of incorporating web 2.0 platforms into mainstream institutional service delivery need to be subject to further research, so that the risks associated with these innovations are better understood at the strategic and policy-making level. Practical implications - This paper makes some recommendations about new principles of library and information practice which will help practitioners make better sense of these innovations in their overall information environment. Social implications - The paper puts in context some of the more problematic social impacts of web 2.0 innovations, without denying the undeniable positive contribution of social networking to the sphere of human interactivity. Originality/value - This paper raises some cautionary points about web 2.0 applications without adopting a precautionary approach of total prohibition. However, none of the suggestions or analysis in this piece should be considered to constitute legal advice. If such advice is required, the reader should consult appropriate legal professionals.
    Date
    22. 1.2011 17:54:04
  19. Luo, Z.; Yu, Y.; Osborne, M.; Wang, T.: Structuring tweets for improving Twitter search (2015) 0.00
    0.0025179111 = product of:
      0.035250753 = sum of:
        0.035250753 = product of:
          0.07050151 = sum of:
            0.07050151 = weight(_text_:texts in 2335) [ClassicSimilarity], result of:
              0.07050151 = score(doc=2335,freq=4.0), product of:
                0.16460659 = queryWeight, product of:
                  5.4822793 = idf(docFreq=499, maxDocs=44218)
                  0.03002521 = queryNorm
                0.42830306 = fieldWeight in 2335, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.4822793 = idf(docFreq=499, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2335)
          0.5 = coord(1/2)
      0.071428575 = coord(1/14)
    
    Abstract
    Spam and wildly varying documents make searching in Twitter challenging. Most Twitter search systems generally treat a Tweet as a plain text when modeling relevance. However, a series of conventions allows users to Tweet in structural ways using a combination of different blocks of texts. These blocks include plain texts, hashtags, links, mentions, etc. Each block encodes a variety of communicative intent and the sequence of these blocks captures changing discourse. Previous work shows that exploiting the structural information can improve the structured documents (e.g., web pages) retrieval. In this study we utilize the structure of Tweets, induced by these blocks, for Twitter retrieval and Twitter opinion retrieval. For Twitter retrieval, a set of features, derived from the blocks of text and their combinations, is used into a learning-to-rank scenario. We show that structuring Tweets can achieve state-of-the-art performance. Our approach does not rely on social media features, but when we do add this additional information, performance improves significantly. For Twitter opinion retrieval, we explore the question of whether structural information derived from the body of Tweets and opinionatedness ratings of Tweets can improve performance. Experimental results show that retrieval using a novel unsupervised opinionatedness feature based on structuring Tweets achieves comparable performance with a supervised method using manually tagged Tweets. Topic-related specific structured Tweet sets are shown to help with query-dependent opinion retrieval.
  20. Mahesh, K.; Karanth, P.: ¬A novel knowledge organization scheme for the Web : superlinks with semantic roles (2012) 0.00
    0.0023989913 = product of:
      0.033585876 = sum of:
        0.033585876 = product of:
          0.06717175 = sum of:
            0.06717175 = weight(_text_:schemes in 822) [ClassicSimilarity], result of:
              0.06717175 = score(doc=822,freq=4.0), product of:
                0.16067243 = queryWeight, product of:
                  5.3512506 = idf(docFreq=569, maxDocs=44218)
                  0.03002521 = queryNorm
                0.41806644 = fieldWeight in 822, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.3512506 = idf(docFreq=569, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=822)
          0.5 = coord(1/2)
      0.071428575 = coord(1/14)
    
    Abstract
    We discuss the needs of a knowledge organization scheme for supporting Web-based software applications. We show how it differs from traditional knowledge organization schemes due to the virtual, dynamic, ad-hoc, userspecific and application-specific nature of Web-based knowledge. The sheer size of Web resources also adds to the complexity of organizing knowledge on the Web. As such, a standard, global scheme such as a single ontology for classifying and organizing all Web-based content is unrealistic. There is nevertheless a strong and immediate need for effective knowledge organization schemes to improve the efficiency and effectiveness of Web-based applications. In this context, we propose a novel knowledge organization scheme wherein concepts in the ontology of a domain are semantically interlinked with specific pieces of Web-based content using a rich hyper-linking structure known as Superlinks with well-defined semantic roles. We illustrate how such a knowledge organization scheme improves the efficiency and effectiveness of a Web-based e-commerce retail store.

Languages

  • e 34
  • d 9

Types

  • a 35
  • el 6
  • m 5
  • s 1
  • More… Less…