Search (58 results, page 1 of 3)

  • × theme_ss:"Internet"
  • × year_i:[2010 TO 2020}
  1. Song, L.; Tso, G.; Fu, Y.: Click behavior and link prioritization : multiple demand theory application for web improvement (2019) 0.01
    0.006445084 = product of:
      0.025780335 = sum of:
        0.017352559 = product of:
          0.052057672 = sum of:
            0.052057672 = weight(_text_:problem in 5322) [ClassicSimilarity], result of:
              0.052057672 = score(doc=5322,freq=4.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.39792046 = fieldWeight in 5322, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5322)
          0.33333334 = coord(1/3)
        0.008427775 = product of:
          0.025283325 = sum of:
            0.025283325 = weight(_text_:29 in 5322) [ClassicSimilarity], result of:
              0.025283325 = score(doc=5322,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.23319192 = fieldWeight in 5322, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5322)
          0.33333334 = coord(1/3)
      0.25 = coord(2/8)
    
    Abstract
    A common problem encountered in Web improvement is how to arrange the homepage links of a Website. This study analyses Web information search behavior, and applies the multiple demand theory to propose two models to help a visitor allocate time for multiple links. The process of searching is viewed as a formal choice problem in which the visitor attempts to choose from multiple Web links to maximize the total utility. The proposed models are calibrated to clickstream data collected from an educational institute over a seven-and-a-half month period. Based on the best fit model, a metric, utility loss, is constructed to measure the performance of each link and arrange them accordingly. Empirical results show that the proposed metric is highly efficient for prioritizing the links on a homepage and the methodology can also be used to study the feasibility of introducing a new function in a Website.
    Date
    6. 7.2019 19:37:29
  2. Landwehr, A.: China schafft digitales Punktesystem für den "besseren" Menschen (2018) 0.01
    0.0055932454 = product of:
      0.044745963 = sum of:
        0.044745963 = product of:
          0.06711894 = sum of:
            0.033711098 = weight(_text_:29 in 4314) [ClassicSimilarity], result of:
              0.033711098 = score(doc=4314,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.31092256 = fieldWeight in 4314, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4314)
            0.03340785 = weight(_text_:22 in 4314) [ClassicSimilarity], result of:
              0.03340785 = score(doc=4314,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.30952093 = fieldWeight in 4314, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4314)
          0.6666667 = coord(2/3)
      0.125 = coord(1/8)
    
    Date
    22. 6.2018 14:29:46
  3. Andrade, T.C.; Dodebei, V.: Traces of digitized newspapers and bom-digital news sites : a trail to the memory on the internet (2016) 0.01
    0.0055932454 = product of:
      0.044745963 = sum of:
        0.044745963 = product of:
          0.06711894 = sum of:
            0.033711098 = weight(_text_:29 in 4901) [ClassicSimilarity], result of:
              0.033711098 = score(doc=4901,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.31092256 = fieldWeight in 4901, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4901)
            0.03340785 = weight(_text_:22 in 4901) [ClassicSimilarity], result of:
              0.03340785 = score(doc=4901,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.30952093 = fieldWeight in 4901, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4901)
          0.6666667 = coord(2/3)
      0.125 = coord(1/8)
    
    Date
    19. 1.2019 17:42:22
    Source
    Knowledge organization for a sustainable world: challenges and perspectives for cultural, scientific, and technological sharing in a connected society : proceedings of the Fourteenth International ISKO Conference 27-29 September 2016, Rio de Janeiro, Brazil / organized by International Society for Knowledge Organization (ISKO), ISKO-Brazil, São Paulo State University ; edited by José Augusto Chaves Guimarães, Suellen Oliveira Milani, Vera Dodebei
  4. Bhatia, S.; Biyani, P.; Mitra, P.: Identifying the role of individual user messages in an online discussion and its use in thread retrieval (2016) 0.01
    0.005355108 = product of:
      0.021420432 = sum of:
        0.014460463 = product of:
          0.04338139 = sum of:
            0.04338139 = weight(_text_:problem in 2650) [ClassicSimilarity], result of:
              0.04338139 = score(doc=2650,freq=4.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.33160037 = fieldWeight in 2650, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2650)
          0.33333334 = coord(1/3)
        0.0069599687 = product of:
          0.020879906 = sum of:
            0.020879906 = weight(_text_:22 in 2650) [ClassicSimilarity], result of:
              0.020879906 = score(doc=2650,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.19345059 = fieldWeight in 2650, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2650)
          0.33333334 = coord(1/3)
      0.25 = coord(2/8)
    
    Abstract
    Online discussion forums have become a popular medium for users to discuss with and seek information from other users having similar interests. A typical discussion thread consists of a sequence of posts posted by multiple users. Each post in a thread serves a different purpose providing different types of information and, thus, may not be equally useful for all applications. Identifying the purpose and nature of each post in a discussion thread is thus an interesting research problem as it can help in improving information extraction and intelligent assistance techniques. We study the problem of classifying a given post as per its purpose in the discussion thread and employ features based on the post's content, structure of the thread, behavior of the participating users, and sentiment analysis of the post's content. We evaluate our approach on two forum data sets belonging to different genres and achieve strong classification performance. We also analyze the relative importance of different features used for the post classification task. Next, as a use case, we describe how the post class information can help in thread retrieval by incorporating this information in a state-of-the-art thread retrieval model.
    Date
    22. 1.2016 11:50:46
  5. Hartmann, B.: Ab ins MoMA : zum virtuellen Museumsgang (2011) 0.01
    0.005155518 = product of:
      0.020622073 = sum of:
        0.012270111 = product of:
          0.03681033 = sum of:
            0.03681033 = weight(_text_:problem in 1821) [ClassicSimilarity], result of:
              0.03681033 = score(doc=1821,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.28137225 = fieldWeight in 1821, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1821)
          0.33333334 = coord(1/3)
        0.008351962 = product of:
          0.025055885 = sum of:
            0.025055885 = weight(_text_:22 in 1821) [ClassicSimilarity], result of:
              0.025055885 = score(doc=1821,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.23214069 = fieldWeight in 1821, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1821)
          0.33333334 = coord(1/3)
      0.25 = coord(2/8)
    
    Content
    "Bin gestern im MoMA gewesen. Hab dann noch mal in der Tate vorbeigeschaut. Nachher dann noch einen Abstecher in die Alte Nationalgalerie gemacht. New York, London, Berlin an einem Tag, das ist dank Google kein Problem mehr. Auf der Plattform www.googleartproject.com bietet der Netz-Gigant jetzt einige virtuelle Museumsrundgänge durch einige der bekanntesten und bedeutendsten Häuser der Welt an. Googles neuestes Angebot unterscheidet sich von den in der Regel gut aufgestellten Homepages der Museen vor allem durch die Street-View-Technologie, mit der man von Raum zu Raum und von Bild zu Bild wandeln kann. Dazu hat der Besucher die Möglichkeit, zahlreiche Informationen zu den Kunstwerken abzurufen, oder die Werke in hochauflösenden Vergrößerungen anzuschauen. Aus jeder der teilnehmenden 17 Sammlungen hat Google ein Werk in einer Auflösung von sieben Milliarden Pixeln fotografiert. Wenn man da beim Rundgang durch die Alte Nationalgalerie auf Edouard Manets "Dans la Serre" (Im Wintergarten) trifft und nur einmal den Fingerring des Mannes heranzoomt, wird er gleichsam zur Unkenntlichkeit vergrößert, und man sieht wie unter einem Mikroskop jedes kleinste Detail des Farbauftrags, jeden feinsten Riss. Faszinierend. Auch wenn es das Original nicht ersetzt."
    Date
    3. 5.1997 8:44:22
  6. Wang, C.; Zhao, S.; Kalra, A.; Borcea, C.; Chen, Y.: Predictive models and analysis for webpage depth-level dwell time (2018) 0.00
    0.0043120594 = product of:
      0.017248238 = sum of:
        0.010225092 = product of:
          0.030675275 = sum of:
            0.030675275 = weight(_text_:problem in 4370) [ClassicSimilarity], result of:
              0.030675275 = score(doc=4370,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.23447686 = fieldWeight in 4370, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4370)
          0.33333334 = coord(1/3)
        0.007023146 = product of:
          0.021069437 = sum of:
            0.021069437 = weight(_text_:29 in 4370) [ClassicSimilarity], result of:
              0.021069437 = score(doc=4370,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.19432661 = fieldWeight in 4370, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4370)
          0.33333334 = coord(1/3)
      0.25 = coord(2/8)
    
    Abstract
    A half of online display ads are not rendered viewable because the users do not scroll deep enough or spend sufficient time at the page depth where the ads are placed. In order to increase the marketing efficiency and ad effectiveness, there is a strong demand for viewability prediction from both advertisers and publishers. This paper aims to predict the dwell time for a given urn:x-wiley:23301635:media:asi24025:asi24025-math-0001 triplet based on historic data collected by publishers. This problem is difficult because of user behavior variability and data sparsity. To solve it, we propose predictive models based on Factorization Machines and Field-aware Factorization Machines in order to overcome the data sparsity issue and provide flexibility to add auxiliary information such as the visible area of a user's browser. In addition, we leverage the prior dwell time behavior of the user within the current page view, that is, time series information, to further improve the proposed models. Experimental results using data from a large web publisher demonstrate that the proposed models outperform comparison models. Also, the results show that adding time series information further improves the performance.
    Date
    29. 9.2018 11:32:23
  7. Kaeser, E.: ¬Das postfaktische Zeitalter (2016) 0.00
    0.0025299042 = product of:
      0.020239234 = sum of:
        0.020239234 = product of:
          0.030358851 = sum of:
            0.012641663 = weight(_text_:29 in 3080) [ClassicSimilarity], result of:
              0.012641663 = score(doc=3080,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.11659596 = fieldWeight in 3080, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=3080)
            0.017717188 = weight(_text_:22 in 3080) [ClassicSimilarity], result of:
              0.017717188 = score(doc=3080,freq=4.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.16414827 = fieldWeight in 3080, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=3080)
          0.6666667 = coord(2/3)
      0.125 = coord(1/8)
    
    Content
    "Es gibt Daten, Informationen und Fakten. Wenn man mir eine Zahlenreihe vorsetzt, dann handelt es sich um Daten: unterscheidbare Einheiten, im Fachjargon: Items. Wenn man mir sagt, dass diese Items stündliche Temperaturangaben der Aare im Berner Marzilibad bedeuten, dann verfüge ich über Information - über interpretierte Daten. Wenn man mir sagt, dies seien die gemessenen Aaretemperaturen am 22. August 2016 im Marzili, dann ist das ein Faktum: empirisch geprüfte interpretierte Daten. Dieser Dreischritt - Unterscheiden, Interpretieren, Prüfen - bildet quasi das Bindemittel des Faktischen, «the matter of fact». Wir alle führen den Dreischritt ständig aus und gelangen so zu einem relativ verlässlichen Wissen und Urteilsvermögen betreffend die Dinge des Alltags. Aber wie schon die Kurzcharakterisierung durchblicken lässt, bilden Fakten nicht den Felsengrund der Realität. Sie sind kritikanfällig, sowohl von der Interpretation wie auch von der Prüfung her gesehen. Um bei unserem Beispiel zu bleiben: Es kann durchaus sein, dass man uns zwei unterschiedliche «faktische» Temperaturverläufe der Aare am 22. August 2016 vorsetzt.
    Date
    24. 8.2016 9:29:24
  8. Pereira, D.A.; Ribeiro-Neto, B.; Ziviani, N.; Laender, A.H.F.; Gonçalves, M.A.: ¬A generic Web-based entity resolution framework (2011) 0.00
    0.0022137975 = product of:
      0.01771038 = sum of:
        0.01771038 = product of:
          0.05313114 = sum of:
            0.05313114 = weight(_text_:problem in 4450) [ClassicSimilarity], result of:
              0.05313114 = score(doc=4450,freq=6.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.4061259 = fieldWeight in 4450, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4450)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    Web data repositories usually contain references to thousands of real-world entities from multiple sources. It is not uncommon that multiple entities share the same label (polysemes) and that distinct label variations are associated with the same entity (synonyms), which frequently leads to ambiguous interpretations. Further, spelling variants, acronyms, abbreviated forms, and misspellings compound to worsen the problem. Solving this problem requires identifying which labels correspond to the same real-world entity, a process known as entity resolution. One approach to solve the entity resolution problem is to associate an authority identifier and a list of variant forms with each entity-a data structure known as an authority file. In this work, we propose a generic framework for implementing a method for generating authority files. Our method uses information from the Web to improve the quality of the authority file and, because of that, is referred to as WER-Web-based Entity Resolution. Our contribution here is threefold: (a) we discuss how to implement the WER framework, which is flexible and easy to adapt to new domains; (b) we run extended experimentation with our WER framework to show that it outperforms selected baselines; and (c) we compare the results of a specialized solution for author name resolution with those produced by the generic WER framework, and show that the WER results remain competitive.
  9. Hwang, S.-Y.; Yang, W.-S.; Ting, K.-D.: Automatic index construction for multimedia digital libraries (2010) 0.00
    0.0021690698 = product of:
      0.017352559 = sum of:
        0.017352559 = product of:
          0.052057672 = sum of:
            0.052057672 = weight(_text_:problem in 4228) [ClassicSimilarity], result of:
              0.052057672 = score(doc=4228,freq=4.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.39792046 = fieldWeight in 4228, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4228)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    Indexing remains one of the most popular tools provided by digital libraries to help users identify and understand the characteristics of the information they need. Despite extensive studies of the problem of automatic index construction for text-based digital libraries, the construction of multimedia digital libraries continues to represent a challenge, because multimedia objects usually lack sufficient text information to ensure reliable index learning. This research attempts to tackle the problem of automatic index construction for multimedia objects by employing Web usage logs and limited keywords pertaining to multimedia objects. The tests of two proposed algorithms use two different data sets with different amounts of textual information. Web usage logs offer precious information for building indexes of multimedia digital libraries with limited textual information. The proposed methods generally yield better indexes, especially for the artwork data set.
  10. Schultz, S.: ¬Die eine App für alles : Mobile Zukunft in China (2016) 0.00
    0.0019685763 = product of:
      0.01574861 = sum of:
        0.01574861 = product of:
          0.04724583 = sum of:
            0.04724583 = weight(_text_:22 in 4313) [ClassicSimilarity], result of:
              0.04724583 = score(doc=4313,freq=4.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.4377287 = fieldWeight in 4313, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4313)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    22. 6.2018 14:22:02
  11. Spink, A.; Du, J.T.: Toward a Web search model : integrating multitasking, cognitive coordination, and cognitive shifts (2011) 0.00
    0.0018075579 = product of:
      0.014460463 = sum of:
        0.014460463 = product of:
          0.04338139 = sum of:
            0.04338139 = weight(_text_:problem in 4624) [ClassicSimilarity], result of:
              0.04338139 = score(doc=4624,freq=4.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.33160037 = fieldWeight in 4624, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4624)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    Limited research has investigated the role of multitasking, cognitive coordination, and cognitive shifts during web search. Understanding these three behaviors is crucial to web search model development. This study aims to explore characteristics of multitasking behavior, types of cognitive shifts, and levels of cognitive coordination as well as the relationship between them during web search. Data collection included pre- and postquestionnaires, think-aloud protocols, web search logs, observations, and interviews with 42 graduate students who conducted 315 web search sessions with 221 information problems. Results show that web search is a dynamic interaction including the ordering of multiple information problems and the generation of evolving information problems, including task switching, multitasking, explicit task and implicit mental coordination, and cognitive shifting. Findings show that explicit task-level coordination is closely linked to multitasking, and implicit cognitive-level coordination is related to the task-coordination process; including information problem development and task switching. Coordination mechanisms directly result in cognitive state shifts including strategy, evaluation, and view states that affect users' holistic shifts in information problem understanding and knowledge contribution. A web search model integrating multitasking, cognitive coordination, and cognitive shifts (MCC model) is presented. Implications and further research also are discussed.
  12. Aksoy, C.; Can, F.; Kocberber, S.: Novelty detection for topic tracking (2012) 0.00
    0.0018075579 = product of:
      0.014460463 = sum of:
        0.014460463 = product of:
          0.04338139 = sum of:
            0.04338139 = weight(_text_:problem in 51) [ClassicSimilarity], result of:
              0.04338139 = score(doc=51,freq=4.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.33160037 = fieldWeight in 51, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=51)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    Multisource web news portals provide various advantages such as richness in news content and an opportunity to follow developments from different perspectives. However, in such environments, news variety and quantity can have an overwhelming effect. New-event detection and topic-tracking studies address this problem. They examine news streams and organize stories according to their events; however, several tracking stories of an event/topic may contain no new information (i.e., no novelty). We study the novelty detection (ND) problem on the tracking news of a particular topic. For this purpose, we build a Turkish ND test collection called BilNov-2005 and propose the usage of three ND methods: a cosine-similarity (CS)-based method, a language-model (LM)-based method, and a cover-coefficient (CC)-based method. For the LM-based ND method, we show that a simpler smoothing approach, Dirichlet smoothing, can have similar performance to a more complex smoothing approach, Shrinkage smoothing. We introduce a baseline that shows the performance of a system with random novelty decisions. In addition, a category-based threshold learning method is used for the first time in ND literature. The experimental results show that the LM-based ND method significantly outperforms the CS- and CC-based methods, and category-based threshold learning achieves promising results when compared to general threshold learning.
  13. Kong, S.; Ye, F.; Feng, L.; Zhao, Z.: Towards the prediction problems of bursting hashtags on Twitter (2015) 0.00
    0.001789391 = product of:
      0.014315128 = sum of:
        0.014315128 = product of:
          0.042945385 = sum of:
            0.042945385 = weight(_text_:problem in 2338) [ClassicSimilarity], result of:
              0.042945385 = score(doc=2338,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.3282676 = fieldWeight in 2338, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=2338)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    Hundreds of thousands of hashtags are generated every day on Twitter. Only a few will burst and become trending topics. In this article, we provide the definition of a bursting hashtag and conduct a systematic study of a series of challenging prediction problems that span the entire life cycles of bursting hashtags. Around the problem of "how to build a system to predict bursting hashtags," we explore different types of features and present machine learning solutions. On real data sets from Twitter, experiments are conducted to evaluate the effectiveness of the proposed solutions and the contributions of features.
  14. Strittmatter, K.: Chinas digitaler Plan für den besseren Menschen (2017) 0.00
    0.0017557865 = product of:
      0.014046292 = sum of:
        0.014046292 = product of:
          0.042138875 = sum of:
            0.042138875 = weight(_text_:29 in 2114) [ClassicSimilarity], result of:
              0.042138875 = score(doc=2114,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.38865322 = fieldWeight in 2114, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2114)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    19. 9.2017 14:45:29
  15. Chen, Y.-L.; Chuang, C.-H.; Chiu, Y.-T.: Community detection based on social interactions in a social network (2014) 0.00
    0.0015337638 = product of:
      0.012270111 = sum of:
        0.012270111 = product of:
          0.03681033 = sum of:
            0.03681033 = weight(_text_:problem in 1219) [ClassicSimilarity], result of:
              0.03681033 = score(doc=1219,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.28137225 = fieldWeight in 1219, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1219)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    Recent research has involved identifying communities in networks. Traditional methods of community detection usually assume that the network's structural information is fully known, which is not the case in many practical networks. Moreover, most previous community detection algorithms do not differentiate multiple relationships between objects or persons in the real world. In this article, we propose a new approach that utilizes social interaction data (e.g., users' posts on Facebook) to address the community detection problem in Facebook and to find the multiple social groups of a Facebook user. Some advantages to our approach are (a) it does not depend on structural information, (b) it differentiates the various relationships that exist among friends, and (c) it can discover a target user's multiple communities. In the experiment, we detect the community distribution of Facebook users using the proposed method. The experiment shows that our method can achieve the result of having the average scores of Total-Community-Purity and Total-Cluster-Purity both at approximately 0.8.
  16. Dupont, J.: Falsch! (2017) 0.00
    0.0014046291 = product of:
      0.011237033 = sum of:
        0.011237033 = product of:
          0.033711098 = sum of:
            0.033711098 = weight(_text_:29 in 3470) [ClassicSimilarity], result of:
              0.033711098 = score(doc=3470,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.31092256 = fieldWeight in 3470, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3470)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    20.11.2017 11:40:29
  17. Kaden, B.; Kindling, M.: Kommunikation und Kontext : Überlegungen zur Entwicklung virtueller Diskursräume für die Wissenschaft (2010) 0.00
    0.0014046291 = product of:
      0.011237033 = sum of:
        0.011237033 = product of:
          0.033711098 = sum of:
            0.033711098 = weight(_text_:29 in 4271) [ClassicSimilarity], result of:
              0.033711098 = score(doc=4271,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.31092256 = fieldWeight in 4271, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4271)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    2. 2.2011 18:29:45
  18. Sood, S.O.; Churchill, E.F.; Antin, J.: Automatic identification of personal insults on social news sites (2012) 0.00
    0.0012781365 = product of:
      0.010225092 = sum of:
        0.010225092 = product of:
          0.030675275 = sum of:
            0.030675275 = weight(_text_:problem in 4976) [ClassicSimilarity], result of:
              0.030675275 = score(doc=4976,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.23447686 = fieldWeight in 4976, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4976)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    As online communities grow and the volume of user-generated content increases, the need for community management also rises. Community management has three main purposes: to create a positive experience for existing participants, to promote appropriate, socionormative behaviors, and to encourage potential participants to make contributions. Research indicates that the quality of content a potential participant sees on a site is highly influential; off-topic, negative comments with malicious intent are a particularly strong boundary to participation or set the tone for encouraging similar contributions. A problem for community managers, therefore, is the detection and elimination of such undesirable content. As a community grows, this undertaking becomes more daunting. Can an automated system aid community managers in this task? In this paper, we address this question through a machine learning approach to automatic detection of inappropriate negative user contributions. Our training corpus is a set of comments from a news commenting site that we tasked Amazon Mechanical Turk workers with labeling. Each comment is labeled for the presence of profanity, insults, and the object of the insults. Support vector machines trained on these data are combined with relevance and valence analysis systems in a multistep approach to the detection of inappropriate negative user contributions. The system shows great potential for semiautomated community management.
  19. Keikha, M.; Crestani, F.; Carman, M.J.: Employing document dependency in blog search (2012) 0.00
    0.0012781365 = product of:
      0.010225092 = sum of:
        0.010225092 = product of:
          0.030675275 = sum of:
            0.030675275 = weight(_text_:problem in 4987) [ClassicSimilarity], result of:
              0.030675275 = score(doc=4987,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.23447686 = fieldWeight in 4987, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4987)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    The goal in blog search is to rank blogs according to their recurrent relevance to the topic of the query. State-of-the-art approaches view it as an expert search or resource selection problem. We investigate the effect of content-based similarity between posts on the performance of the retrieval system. We test two different approaches for smoothing (regularizing) relevance scores of posts based on their dependencies. In the first approach, we smooth term distributions describing posts by performing a random walk over a document-term graph in which similar posts are highly connected. In the second, we directly smooth scores for posts using a regularization framework that aims to minimize the discrepancy between scores for similar documents. We then extend these approaches to consider the time interval between the posts in smoothing the scores. The idea is that if two posts are temporally close, then they are good sources for smoothing each other's relevance scores. We compare these methods with the state-of-the-art approaches in blog search that employ Language Modeling-based resource selection algorithms and fusion-based methods for aggregating post relevance scores. We show performance gains over the baseline techniques which do not take advantage of the relation between posts for smoothing relevance estimates.
  20. Paltoglou, G.: Sentiment-based event detection in Twitter (2016) 0.00
    0.0012781365 = product of:
      0.010225092 = sum of:
        0.010225092 = product of:
          0.030675275 = sum of:
            0.030675275 = weight(_text_:problem in 3010) [ClassicSimilarity], result of:
              0.030675275 = score(doc=3010,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.23447686 = fieldWeight in 3010, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3010)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    The main focus of this article is to examine whether sentiment analysis can be successfully used for "event detection," that is, detecting significant events that occur in the world. Most solutions to this problem are typically based on increases or spikes in frequency of terms in social media. In our case, we explore whether sudden changes in the positivity or negativity that keywords are typically associated with can be exploited for this purpose. A data set that contains several million Twitter messages over a 1-month time span is presented and experimental results demonstrate that sentiment analysis can be successfully utilized for this purpose. Further experiments study the sensitivity of both frequency- or sentiment-based solutions to a number of parameters. Concretely, we show that the number of tweets that are used for event detection is an important factor, while the number of days used to extract token frequency or sentiment averages is not. Lastly, we present results focusing on detecting local events and conclude that all approaches are dependant on the level of coverage that such events receive in social media.

Languages

  • e 40
  • d 17

Types

  • a 52
  • el 6
  • m 4
  • s 1
  • More… Less…