Diese Datenbank enthält über 40.000 Dokumente zu Themen aus den Bereichen Formalerschließung – Inhaltserschließung – Information Retrieval.
© 2015 W. Gödert, TH Köln, Institut für Informationswissenschaft / Powered by litecat, BIS Oldenburg (Stand: 04. Juni 2021)
1Nicholson, S. ; Smith, C.A.: Using lessons from health care to protect the privacy of library users : guidelines for the de-identification of library data based on HIPAA.
In: Journal of the American Society for Information Science and Technology. 58(2007) no.8, S.1198-1206.
Abstract: Although libraries have employed policies to protect the data about use of their services, these policies are rarely specific or standardized. Since 1996, the U.S. health care system has been grappling with the Health Insurance Portability and Accountability Act (HIPAA; Health Insurance Portability and Accountability Act, 1996), which is designed to provide those handling personal health information with standardized, definitive instructions as to the protection of data. In this work, the authors briefly discuss the present situation of privacy policies about library use data, outline the HIPAA guidelines to understand parallels between the two, and finally propose methods to create a de-identified library data warehouse based on HIPAA for the protection of user privacy.
2Nicholson, S. ; Sierra, T. ; Eseryel, U.Y. ; Park, J.-H. ; Barkow, P. ; Pozo, E.J. ; Ward, J.: How much of it is real? : analysis of paid placement in Web search engine results.
In: Journal of the American Society for Information Science and Technology. 57(2006) no.4, S.448-461.
Abstract: Most Web search tools integrate sponsored results with results from their internal editorial database in providing results to users. The goal of this research is to get a better idea of how much of the screen real estate displays real editorial results as compared to sponsored results. The overall average results are that 40% of all results presented on the first screen are real results, and when the entire first Web page is considered, 67% of the results are nonsponsored results. For general search tools such as Google, 56% of the first screen and 82% of the first Web page contain nonsponsored results. Other results include that query structure makes a significant difference in the percentage of nonsponsored results returned by a search. Similarly, the topic of the query also can have a significant effect on the percentage of sponsored results displayed by most Web search tools.
3Pomerantz, J. ; Nicholson, S. ; Belanger, Y. ; Lankes, R.D.: ¬The current state of digital reference : validation of a general digital reference model through a survey of digital reference services.
In: Information processing and management. 40(2004) no.2, S.347-363.
Abstract: This paper describes a study conducted to determine the paths digital reference services take through a general process model of asynchronous digital reference. A survey based on the general process model was conducted; each decision point in this model provided the basis for at least one question. Common, uncommon, and wished-for practices are identified, as well as correlations between characteristics of services and the practices employed by those services. Identification of such trends has implications for the development of software tools for digital reference. This study presents a snapshot of the state-of-the-art in digital reference as of late 2001- early 2002, and validates the general process model of asynchronous digital reference.
Themenfeld: Informationsdienstleistungen ; Internet
4Nicholson, S.: Bibliomining for automated collection development in a digital library setting : using data mining to discover Web-based scholarly research works.
In: Journal of the American Society for Information Science and technology. 54(2003) no.12, S.1081-1090.
Abstract: This research creates an intelligent agent for automated collection development in a digital library setting. It uses a predictive model based an facets of each Web page to select scholarly works. The criteria came from the academic library selection literature, and a Delphi study was used to refine the list to 41 criteria. A Perl program was designed to analyze a Web page for each criterion and applied to a large collection of scholarly and nonscholarly Web pages. Bibliomining, or data mining for libraries, was then used to create different classification models. Four techniques were used: logistic regression, nonparametric discriminant analysis, classification trees, and neural networks. Accuracy and return were used to judge the effectiveness of each model an test datasets. In addition, a set of problematic pages that were difficult to classify because of their similarity to scholarly research was gathered and classified using the models. The resulting models could be used in the selection process to automatically create a digital library of Webbased scholarly research works. In addition, the technique can be extended to create a digital library of any type of structured electronic information.
Themenfeld: Data Mining
5Nicholson, S.: Raising reliability of Web search tool research through replication and chaos theory.
In: Journal of the American Society for Information Science. 51(2000) no.8, S.724-729.
Abstract: Because the WWW is a dynamic collection of information, the Web search tools (or 'search engines') that index the Web are dynamic. Traditional information retrieval evaluation techniques may not provide reliable results when applied to the Web search tools. This study is the result of 10 replications of the 1996 classic Ding and Marchionini Web search tool research. It explores the effects that replication can have on transforming unreliable results from one iteration into replicable and therefore reliable results following multile iterations
Anmerkung: Vgl.: Ding, W. u. G. Marchionini: A comparative study of Web search service performance
Objekt: Infoseek ; Lycos ; AltaVista ; Excite
6Nicholson, S.: ¬A proposal for categorization and nomenclature for Web search tools.
In: Journal of Internet cataloging. 2(2000) nos.3/4, S.9-28.
Abstract: Ambiguities in Web search tool (more commonly known as "search engine") terminology are problematic when conducting precise, replicable research or when teaching others to use search tools. Standardized terminology would enable Web searchers to be aware of subtle differences between Web search tools and the implications of these for searching. A categorization and nomenclature for standardized classifications of different aspects of Web search tools is proposed, and advantages and disadvantages of using tools in each category are discussed