-
TREC: experiment and evaluation in information retrieval (2005)
0.02
0.018613702 = sum of:
0.0074650636 = product of:
0.029860254 = sum of:
0.029860254 = weight(_text_:authors in 636) [ClassicSimilarity], result of:
0.029860254 = score(doc=636,freq=2.0), product of:
0.2371355 = queryWeight, product of:
4.558814 = idf(docFreq=1258, maxDocs=44218)
0.05201693 = queryNorm
0.12592064 = fieldWeight in 636, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
4.558814 = idf(docFreq=1258, maxDocs=44218)
0.01953125 = fieldNorm(doc=636)
0.25 = coord(1/4)
0.011148638 = product of:
0.022297276 = sum of:
0.022297276 = weight(_text_:t in 636) [ClassicSimilarity], result of:
0.022297276 = score(doc=636,freq=2.0), product of:
0.20491594 = queryWeight, product of:
3.9394085 = idf(docFreq=2338, maxDocs=44218)
0.05201693 = queryNorm
0.10881182 = fieldWeight in 636, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.9394085 = idf(docFreq=2338, maxDocs=44218)
0.01953125 = fieldNorm(doc=636)
0.5 = coord(1/2)
- Content
- Enthält die Beiträge: 1. The Text REtrieval Conference - Ellen M. Voorhees and Donna K. Harman 2. The TREC Test Collections - Donna K. Harman 3. Retrieval System Evaluation - Chris Buckley and Ellen M. Voorhees 4. The TREC Ad Hoc Experiments - Donna K. Harman 5. Routing and Filtering - Stephen Robertson and Jamie Callan 6. The TREC Interactive Tracks: Putting the User into Search - Susan T. Dumais and Nicholas J. Belkin 7. Beyond English - Donna K. Harman 8. Retrieving Noisy Text - Ellen M. Voorhees and John S. Garofolo 9.The Very Large Collection and Web Tracks - David Hawking and Nick Craswell 10. Question Answering in TREC - Ellen M. Voorhees 11. The University of Massachusetts and a Dozen TRECs - James Allan, W. Bruce Croft and Jamie Callan 12. How Okapi Came to TREC - Stephen Robertson 13. The SMART Project at TREC - Chris Buckley 14. Ten Years of Ad Hoc Retrieval at TREC Using PIRCS - Kui-Lam Kwok 15. MultiText Experiments for TREC - Gordon V. Cormack, Charles L. A. Clarke, Christopher R. Palmer and Thomas R. Lynam 16. A Language-Modeling Approach to TREC - Djoerd Hiemstra and Wessel Kraaij 17. BM Research Activities at TREC - Eric W. Brown, David Carmel, Martin Franz, Abraham Ittycheriah, Tapas Kanungo, Yoelle Maarek, J. Scott McCarley, Robert L. Mack, John M. Prager, John R. Smith, Aya Soffer, Jason Y. Zien and Alan D. Marwick Epilogue: Metareflections on TREC - Karen Sparck Jones
- Footnote
- ... TREC: Experiment and Evaluation in Information Retrieval is a reliable and comprehensive review of the TREC program and has been adopted by NIST as the official history of TREC (see http://trec.nist.gov). We were favorably surprised by the book. Well structured and written, chapters are self-contained and the existence of references to specialized and more detailed publications is continuous, which makes it easier to expand into the different aspects analyzed in the text. This book succeeds in compiling TREC evolution from its inception in 1992 to 2003 in an adequate and manageable volume. Thanks to the impressive effort performed by the authors and their experience in the field, it can satiate the interests of a great variety of readers. While expert researchers in the IR field and IR-related industrial companies can use it as a reference manual, it seems especially useful for students and non-expert readers willing to approach this research area. Like NIST, we would recommend this reading to anyone who may be interested in textual information retrieval."
-
O'Connor, B.C.; Kearns, J.; Anderson, R.L.: Doing things with information : beyond indexing and abstracting (2008)
0.01
0.011944101 = product of:
0.023888202 = sum of:
0.023888202 = product of:
0.09555281 = sum of:
0.09555281 = weight(_text_:authors in 4297) [ClassicSimilarity], result of:
0.09555281 = score(doc=4297,freq=8.0), product of:
0.2371355 = queryWeight, product of:
4.558814 = idf(docFreq=1258, maxDocs=44218)
0.05201693 = queryNorm
0.40294603 = fieldWeight in 4297, product of:
2.828427 = tf(freq=8.0), with freq of:
8.0 = termFreq=8.0
4.558814 = idf(docFreq=1258, maxDocs=44218)
0.03125 = fieldNorm(doc=4297)
0.25 = coord(1/4)
0.5 = coord(1/2)
- Footnote
- The authors state that this book emerged from a proposal to do a second edition of Explorations in Indexing and Abstracting (O'Connor 1996); much of its content is the result of the authors' reaction to the reviews of this first edition and their realization for "the necessity to address some more fundamental questions". Rez. in: KO 38(2011) no.1, S.62-64 (L.F. Spiteri): "This book provides a good overview of the relationship between the document and the user; in this regard, it reinforces the importance of the clientcentred approach to the design of document representation systems. In the final chapter, the authors state: "We have offered examples of new ways to think about messages in all sorts of media and how they might be discovered, analyzed, synthesized, and generated. We brought together philosophical, scientific, and engineering notions into a fundamental model for just how we might understand doing this with information" (p. 225). The authors have certainly succeeded in highlighting the complex processes, nature, and implications of document representation systems, although, as has been seen, the novelty of some of their discussions and suggestions is sometimes limited. With further explanation, the FOC model may serve as a useful way to understand how to build document representation systems to better meet user needs."; vgl.: http://www.ergon-verlag.de/isko_ko/downloads/ko_38_2011_1e.pdf.
-
Antoniou, G.; Harmelen, F. van: ¬A semantic Web primer (2004)
0.01
0.0074650636 = product of:
0.014930127 = sum of:
0.014930127 = product of:
0.05972051 = sum of:
0.05972051 = weight(_text_:authors in 468) [ClassicSimilarity], result of:
0.05972051 = score(doc=468,freq=8.0), product of:
0.2371355 = queryWeight, product of:
4.558814 = idf(docFreq=1258, maxDocs=44218)
0.05201693 = queryNorm
0.25184128 = fieldWeight in 468, product of:
2.828427 = tf(freq=8.0), with freq of:
8.0 = termFreq=8.0
4.558814 = idf(docFreq=1258, maxDocs=44218)
0.01953125 = fieldNorm(doc=468)
0.25 = coord(1/4)
0.5 = coord(1/2)
- Footnote
- Rez. in: JASIST 57(2006) no.8, S.1132-1133 (H. Che): "The World Wide Web has been the main source of an important shift in the way people communicate with each other, get information, and conduct business. However, most of the current Web content is only suitable for human consumption. The main obstacle to providing better quality of service is that the meaning of Web content is not machine-accessible. The "Semantic Web" is envisioned by Tim Berners-Lee as a logical extension to the current Web that enables explicit representations of term meaning. It aims to bring the Web to its full potential via the exploration of these machine-processable metadata. To fulfill this, it pros ides some meta languages like RDF, OWL, DAML+OIL, and SHOE for expressing knowledge that has clear, unambiguous meanings. The first steps in searing the Semantic Web into the current Web are successfully underway. In the forthcoming years, these efforts still remain highly focused in the research and development community. In the next phase, the Semantic Web will respond more intelligently to user queries. The first chapter gets started with an excellent introduction to the Semantic Web vision. At first, today's Web is introduced, and problems with some current applications like search engines are also covered. Subsequently, knowledge management. business-to-consumer electronic commerce, business-to-business electronic commerce, and personal agents are used as examples to show the potential requirements for the Semantic Web. Next comes the brief description of the underpinning technologies, including metadata, ontology, logic, and agent. The differences between the Semantic Web and Artificial Intelligence are also discussed in a later subsection. In section 1.4, the famous "laser-cake" diagram is given to show a layered view of the Semantic Web. From chapter 2, the book starts addressing some of the most important technologies for constructing the Semantic Web. In chapter 2, the authors discuss XML and its related technologies such as namespaces, XPath, and XSLT. XML is a simple, very flexible text format which is often used for the exchange of a wide variety of data on the Web and elsewhere. The W3C has defined various languages on top of XML, such as RDF. Although this chapter is very well planned and written, many details are not included because of the extensiveness of the XML technologies. Many other books on XML provide more comprehensive coverage.
The next chapter introduces resource description framework (RDF) and RDF schema (RDFS). Unlike XML, RDF provides a foundation for expressing the semantics of dada: it is a standard dada model for machine-processable semantics. Resource description framework schema offers a number of modeling primitives for organizing RDF vocabularies in typed hierarchies. In addition to RDF and RDFS, a query language for RDF, i.e. RQL. is introduced. This chapter and the next chapter are two of the most important chapters in the book. Chapter 4 presents another language called Web Ontology Language (OWL). Because RDFS is quite primitive as a modeling language for the Web, more powerful languages are needed. A richer language. DAML+OIL, is thus proposed as a joint endeavor of the United States and Europe. OWL takes DAML+OIL as the starting point, and aims to be the standardized and broadly accepted ontology language. At the beginning of the chapter, the nontrivial relation with RDF/RDFS is discussed. Then the authors describe the various language elements of OWL in some detail. Moreover, Appendix A contains an abstract OWL syntax. which compresses OWL and makes OWL much easier to read. Chapter 5 covers both monotonic and nonmonotonic rules. Whereas the previous chapter's mainly concentrate on specializations of knowledge representation, this chapter depicts the foundation of knowledge representation and inference. Two examples are also givwn to explain monotonic and non-monotonic rules, respectively. "To get the most out of the chapter. readers had better gain a thorough understanding of predicate logic first. Chapter 6 presents several realistic application scenarios to which the Semantic Web technology can be applied. including horizontal information products at Elsevier, data integration at Audi, skill finding at Swiss Life, a think tank portal at EnerSearch, e-learning. Web services, multimedia collection indexing, online procurement, raid device interoperability. These case studies give us some real feelings about the Semantic Web.
The chapter on ontology engineering describes the development of ontology-based systems for the Web using manual and semiautomatic methods. Ontology is a concept similar to taxonomy. As stated in the introduction, ontology engineering deals with some of the methodological issues that arise when building ontologies, in particular, con-structing ontologies manually, reusing existing ontologies. and using semiautomatic methods. A medium-scale project is included at the end of the chapter. Overall the book is a nice introduction to the key components of the Semantic Web. The reading is quite pleasant, in part due to the concise layout that allows just enough content per page to facilitate readers' comprehension. Furthermore, the book provides a large number of examples, code snippets, exercises, and annotated online materials. Thus, it is very suitable for use as a textbook for undergraduates and low-grade graduates, as the authors say in the preface. However, I believe that not only students but also professionals in both academia and iudustry will benefit from the book. The authors also built an accompanying Web site for the book at http://www.semanticwebprimer.org. On the main page, there are eight tabs for each of the eight chapters. For each tabm the following sections are included: overview, example, presentations, problems and quizzes, errata, and links. These contents will greatly facilitate readers: for example, readers can open the listed links to further their readings. The vacancy of the errata sections also proves the quality of the book."