Cosh, K.J.; Burns, R.; Daniel, T.: Content clouds : classifying content in Web 2.0 (2008)
0.01
0.0063588563 = product of:
0.025435425 = sum of:
0.025435425 = product of:
0.05087085 = sum of:
0.05087085 = weight(_text_:processing in 2013) [ClassicSimilarity], result of:
0.05087085 = score(doc=2013,freq=2.0), product of:
0.18956426 = queryWeight, product of:
4.048147 = idf(docFreq=2097, maxDocs=44218)
0.046827413 = queryNorm
0.26835677 = fieldWeight in 2013, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
4.048147 = idf(docFreq=2097, maxDocs=44218)
0.046875 = fieldNorm(doc=2013)
0.5 = coord(1/2)
0.25 = coord(1/4)
- Abstract
- Purpose - With increasing amounts of user generated content being produced electronically in the form of wikis, blogs, forums etc. the purpose of this paper is to investigate a new approach to classifying ad hoc content. Design/methodology/approach - The approach applies natural language processing (NLP) tools to automatically extract the content of some text, visualizing the results in a content cloud. Findings - Content clouds share the visual simplicity of a tag cloud, but display the details of an article at a different level of abstraction, providing a complimentary classification. Research limitations/implications - Provides the general approach to creating a content cloud. In the future, the process can be refined and enhanced by further evaluation of results. Further work is also required to better identify closely related articles. Practical implications - Being able to automatically classify the content generated by web users will enable others to find more appropriate content. Originality/value - The approach is original. Other researchers have produced a cloud, simply by using skiplists to filter unwanted words, this paper's approach improves this by applying appropriate NLP techniques.