Search (65 results, page 2 of 4)

  • × theme_ss:"Klassifikationssysteme im Online-Retrieval"
  1. Pollitt, A.S.: ¬The key role of classification and indexing in view-based searching (1998) 0.01
    0.011375135 = product of:
      0.034125403 = sum of:
        0.034125403 = product of:
          0.068250805 = sum of:
            0.068250805 = weight(_text_:indexing in 4429) [ClassicSimilarity], result of:
              0.068250805 = score(doc=4429,freq=4.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.3588626 = fieldWeight in 4429, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4429)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    The application of classification schemes and thesauri to improve online information retrieval can be traced back to the beginnings of online searching itself, but the true potential for using knowledge structures in the user interface has yet to be realized. View-based searching seeks to exploit the classified arrangements in thesauri and existing classification schemes to improve the performance of such systems. HIBROWSE for EMBASE is a system which demonstrates the power of applying an approach to information retrieval which is strongly related to faceted classification. It does this by employing a point a click user interface with mutually constraining views utilising knowledge structure hierarchies for both query specification and the presentation of results. The relevance of this approach to library OPACs is discussed in the context of the digital library, concluding that out legacy of research in classification and indexing is more relevant than ever in the design of systems to cope with the problems of information access
  2. Ellis, D.; Vasconcelos, A.: Ranganathan and the Net : using facet analysis to search and organise the World Wide Web (1999) 0.01
    0.011375135 = product of:
      0.034125403 = sum of:
        0.034125403 = product of:
          0.068250805 = sum of:
            0.068250805 = weight(_text_:indexing in 726) [ClassicSimilarity], result of:
              0.068250805 = score(doc=726,freq=4.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.3588626 = fieldWeight in 726, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.046875 = fieldNorm(doc=726)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    This paper documents the continuing relevance of facet analysis as a technique for searching and organising WWW based materials. The 2 approaches underlying WWW searching and indexing - word and concept based indexing - are outlined. It is argued that facet analysis as an a posteriori approach to classification using words from the subject field as the concept terms in the classification derived represents an excellent approach to searching and organising the results of WWW searches using either search engines or search directories. Finally it is argued that the underlying philosophy of facet analysis is better suited to the disparate nature of WWW resources and searchers than the assumptions of contemporaray IR research.
  3. Comaromi, C.L.: Summation of classification as an enhancement of intellectual access to information in an online environment (1990) 0.01
    0.011219318 = product of:
      0.033657953 = sum of:
        0.033657953 = product of:
          0.06731591 = sum of:
            0.06731591 = weight(_text_:22 in 3576) [ClassicSimilarity], result of:
              0.06731591 = score(doc=3576,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.38690117 = fieldWeight in 3576, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3576)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    8. 1.2007 12:22:40
  4. Reiner, U.: Automatische DDC-Klassifizierung von bibliografischen Titeldatensätzen (2009) 0.01
    0.011219318 = product of:
      0.033657953 = sum of:
        0.033657953 = product of:
          0.06731591 = sum of:
            0.06731591 = weight(_text_:22 in 611) [ClassicSimilarity], result of:
              0.06731591 = score(doc=611,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.38690117 = fieldWeight in 611, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=611)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 8.2009 12:54:24
  5. Dunsire, G.: Digital decimals : Dewey and online libraries (2008) 0.01
    0.01072458 = product of:
      0.032173738 = sum of:
        0.032173738 = product of:
          0.064347476 = sum of:
            0.064347476 = weight(_text_:indexing in 2164) [ClassicSimilarity], result of:
              0.064347476 = score(doc=2164,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.3383389 = fieldWeight in 2164, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2164)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Source
    New pespectives on subject indexing and classification: essays in honour of Magda Heiner-Freiling. Red.: K. Knull-Schlomann, u.a
  6. Zaytseva, E.: Classification tools for librarians and users : the Russian experience (2008) 0.01
    0.01072458 = product of:
      0.032173738 = sum of:
        0.032173738 = product of:
          0.064347476 = sum of:
            0.064347476 = weight(_text_:indexing in 2179) [ClassicSimilarity], result of:
              0.064347476 = score(doc=2179,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.3383389 = fieldWeight in 2179, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2179)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Source
    New pespectives on subject indexing and classification: essays in honour of Magda Heiner-Freiling. Red.: K. Knull-Schlomann, u.a
  7. Lösse, M.; Svensson, L.: "Classification at a Crossroad" : Internationales UDC-Seminar 2009 in Den Haag, Niederlande (2010) 0.01
    0.009519908 = product of:
      0.028559722 = sum of:
        0.028559722 = product of:
          0.057119444 = sum of:
            0.057119444 = weight(_text_:22 in 4379) [ClassicSimilarity], result of:
              0.057119444 = score(doc=4379,freq=4.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.32829654 = fieldWeight in 4379, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4379)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Am 29. und 30. Oktober 2009 fand in der Königlichen Bibliothek in Den Haag das zweite internationale UDC-Seminar zum Thema "Classification at a Crossroad" statt. Organisiert wurde diese Konferenz - wie auch die erste Konferenz dieser Art im Jahr 2007 - vom UDC-Konsortium (UDCC). Im Mittelpunkt der diesjährigen Veranstaltung stand die Erschließung des World Wide Web unter besserer Nutzung von Klassifikationen (im Besonderen natürlich der UDC), einschließlich benutzerfreundlicher Repräsentationen von Informationen und Wissen. Standards, neue Technologien und Dienste, semantische Suche und der multilinguale Zugriff spielten ebenfalls eine Rolle. 135 Teilnehmer aus 35 Ländern waren dazu nach Den Haag gekommen. Das Programm umfasste mit 22 Vorträgen aus 14 verschiedenen Ländern eine breite Palette, wobei Großbritannien mit fünf Beiträgen am stärksten vertreten war. Die Tagesschwerpunkte wurden an beiden Konferenztagen durch die Eröffnungsvorträge gesetzt, die dann in insgesamt sechs thematischen Sitzungen weiter vertieft wurden.
    Date
    22. 1.2010 15:06:54
  8. Broughton, V.; Lane, H.: Classification schemes revisited : applications to Web indexing and searching (2000) 0.01
    0.009479279 = product of:
      0.028437834 = sum of:
        0.028437834 = product of:
          0.05687567 = sum of:
            0.05687567 = weight(_text_:indexing in 2476) [ClassicSimilarity], result of:
              0.05687567 = score(doc=2476,freq=4.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.29905218 = fieldWeight in 2476, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2476)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Basic skills of classification and subject indexing have been little taught in British library schools since automation was introduced into libraries. However, development of the Internet as a major medium of publication has stretched the capability of search engines to cope with retrieval. Consequently, there has been interest in applying existing systems of knowledge organization to electronic resources. Unfortunately, the classification systems have been adopted without a full understanding of modern classification principles. Analytico-synthetic schemes have been used crudely, as in the case of the Universal Decimal Classification (UDC). The fully faceted Bliss Bibliographical Classification, 2nd edition (BC2) with its potential as a tool for electronic resource retrieval is virtually unknown outside academic libraries
  9. Slavic, A.: UDC in subject gateways : experiment or opportunity? (2006) 0.01
    0.009479279 = product of:
      0.028437834 = sum of:
        0.028437834 = product of:
          0.05687567 = sum of:
            0.05687567 = weight(_text_:indexing in 4879) [ClassicSimilarity], result of:
              0.05687567 = score(doc=4879,freq=4.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.29905218 = fieldWeight in 4879, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4879)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    The paper gives a short overview of the history of use of UDC in Internet subject gateways (SGs) with an English interface, from 1993 to 2006. There were in total, nine quality controlled SGs that were functional for shorter or longer periods of time. Their typology and functionality is described. Quality SGs have evolved and the role of classification has changed accordingly from supporting subject organization on the interface and automatic categorization of resources, towards supporting a semantic linking, control and vocabulary mapping between different indexing systems in subject hubs and federated SGs. In this period, many SGs ceased to exist and little information remains available regarding their status. SGs currently using UDC, for some part of their resource organization, do not use a UDC subject hierarchy at the interface and its role in resource indexing has become more difficult to observe. Since 2000, UDC has become more prevalent in East European SGs, portals and hubs, which are outside the scope of this research. This paper is an attempt to provide a record on this particular application of UDC and to offer some consideration of the changes in requirements when it comes to the use of library classification in resource discovery.
  10. Lincicum, S.: Critical appraisal of the use of classification in the future : non traditional uses of classification: report of a panel discussion (1995) 0.01
    0.009479279 = product of:
      0.028437834 = sum of:
        0.028437834 = product of:
          0.05687567 = sum of:
            0.05687567 = weight(_text_:indexing in 5570) [ClassicSimilarity], result of:
              0.05687567 = score(doc=5570,freq=4.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.29905218 = fieldWeight in 5570, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5570)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Dagobert Soergel opened the discussion by saying, "Users need our help to find and make sense of information." He said that he believes that classification can provide much of the support users need. He sees little distinction between the concept of a thesaurus and that of classification since both seek to provide structure for knowledge bases in order to facilitate information retrieval. Soergel's discussion of his concept of a multifunctional, multilingual thesaurus comprised the bulk of his presentation. This thesaurus would be a database of concepts, terms, and relationships which would include classification. In this context, classification has a much broader set of functions than it currently does in most American libraries where classification serves primarily as a method of shelf arrangement. The thesaurus Soergel envisions would lay out the semantic map of a field and could therefore be used as a learning tool or as a basis for research planning, or to assist users in clarifying terms and concepts. It could support indexing and searching and provide for the organization of knowledge for expert systems and other artificial intelligence applications. Among its other features, such a thesaurus could assist users in making sense of information by providing structured presentation of search results based on user needs and preferences, and it could enhance natural language processing capabilities such as automated indexing and abstracting and machine translation.
  11. Brandhorst, H.; Huisstede, P.V.: ICONCLASS in de computer : de classificatie van beeldmateriaal in een geautomatiseerde omgeving (1992) 0.01
    0.009384007 = product of:
      0.02815202 = sum of:
        0.02815202 = product of:
          0.05630404 = sum of:
            0.05630404 = weight(_text_:indexing in 6486) [ClassicSimilarity], result of:
              0.05630404 = score(doc=6486,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.29604656 = fieldWeight in 6486, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=6486)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Contribution to thematic issue of art libraries in the Netherlands. Visual images are now widely recognised as an importatnt resource for the historical research. To improve access to collections of such material H. van de Wal developed ICONCLASS in the Netherlands. With some 150.000 controlled indexing terms the scheme is hierarchical and easy to use. In response to requests a version of ICONCLASS has been developed for use in automated systems. The scheme has been used by the Dutch Royal Library for its database of printers' devices, published in CD-ROM version in Dec. 91. Experience with the project shows the need for careful preparations and the use of trained staff
  12. Heiser, W.J.: Zoeken in de bibliografische ruimte : de bijzondere rol van de hierarchische classificaties bij geautomatiseerde onderwerpsontsluiting (1993) 0.01
    0.009384007 = product of:
      0.02815202 = sum of:
        0.02815202 = product of:
          0.05630404 = sum of:
            0.05630404 = weight(_text_:indexing in 3121) [ClassicSimilarity], result of:
              0.05630404 = score(doc=3121,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.29604656 = fieldWeight in 3121, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3121)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Footnote
    Searching in the bibliographical space: the special function of hierarchical classifications in automated subject indexing
  13. Francu, V.: Construirea unui tezaur multilingv bazat pr CZU (1997) 0.01
    0.009384007 = product of:
      0.02815202 = sum of:
        0.02815202 = product of:
          0.05630404 = sum of:
            0.05630404 = weight(_text_:indexing in 3221) [ClassicSimilarity], result of:
              0.05630404 = score(doc=3221,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.29604656 = fieldWeight in 3221, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3221)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    The automation of Bucharest University Central Library involved the compilation of a dictionary of terms to facilitate postcoordinated searching in accordance with the UDC notation attached to every bibliographic record. Describes the project and demonstrates why a multilingual UDC based thesaurus is considered as ideal indexing and searching tool. The experiment, which applied to Class 8 of UDC (Linguisitcs and Literature), illustrates how all UDC tables can be successfully used to build a thesaurus and the ways in which their limitations can be overcome by a thesaurus
  14. Chowdhury, S.; Chowdhury, G.G.: Using DDC to create a visual knowledge map as an aid to online information retrieval (2004) 0.01
    0.009287758 = product of:
      0.027863273 = sum of:
        0.027863273 = product of:
          0.055726547 = sum of:
            0.055726547 = weight(_text_:indexing in 2643) [ClassicSimilarity], result of:
              0.055726547 = score(doc=2643,freq=6.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.2930101 = fieldWeight in 2643, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.03125 = fieldNorm(doc=2643)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Content
    1. Introduction Web search engines and digital libraries usually expect the users to use search terms that most accurately represent their information needs. Finding the most appropriate search terms to represent an information need is an age old problem in information retrieval. Keyword or phrase search may produce good search results as long as the search terms or phrase(s) match those used by the authors and have been chosen for indexing by the concerned information retrieval system. Since this does not always happen, a large number of false drops are produced by information retrieval systems. The retrieval results become worse in very large systems that deal with millions of records, such as the Web search engines and digital libraries. Vocabulary control tools are used to improve the performance of text retrieval systems. Thesauri, the most common type of vocabulary control tool used in information retrieval, appeared in the late fifties, designed for use with the emerging post-coordinate indexing systems of that time. They are used to exert terminology control in indexing, and to aid in searching by allowing the searcher to select appropriate search terms. A large volume of literature exists describing the design features, and experiments with the use, of thesauri in various types of information retrieval systems (see for example, Furnas et.al., 1987; Bates, 1986, 1998; Milstead, 1997, and Shiri et al., 2002).
  15. Doyle, B.: ¬The classification and evaluation of Content Management Systems (2003) 0.01
    0.008975455 = product of:
      0.026926363 = sum of:
        0.026926363 = product of:
          0.053852726 = sum of:
            0.053852726 = weight(_text_:22 in 2871) [ClassicSimilarity], result of:
              0.053852726 = score(doc=2871,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.30952093 = fieldWeight in 2871, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2871)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    30. 7.2004 12:22:52
  16. Peereboom, M.: DutchESS : Dutch Electronic Subject Service - a Dutch national collaborative effort (2000) 0.01
    0.008975455 = product of:
      0.026926363 = sum of:
        0.026926363 = product of:
          0.053852726 = sum of:
            0.053852726 = weight(_text_:22 in 4869) [ClassicSimilarity], result of:
              0.053852726 = score(doc=4869,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.30952093 = fieldWeight in 4869, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4869)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Date
    22. 6.2002 19:39:23
  17. Van Dijck, P.: Introduction to XFML (2003) 0.01
    0.008975455 = product of:
      0.026926363 = sum of:
        0.026926363 = product of:
          0.053852726 = sum of:
            0.053852726 = weight(_text_:22 in 2474) [ClassicSimilarity], result of:
              0.053852726 = score(doc=2474,freq=2.0), product of:
                0.17398734 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.049684696 = queryNorm
                0.30952093 = fieldWeight in 2474, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2474)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Source
    http://www.xml.com/lpt/a/2003/01/22/xfml.html
  18. Poynder, R.: Web research engines? (1996) 0.01
    0.0080434345 = product of:
      0.024130303 = sum of:
        0.024130303 = product of:
          0.048260607 = sum of:
            0.048260607 = weight(_text_:indexing in 5698) [ClassicSimilarity], result of:
              0.048260607 = score(doc=5698,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.2537542 = fieldWeight in 5698, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5698)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Describes the shortcomings of search engines for the WWW comparing their current capabilities to those of the first generation CD-ROM products. Some allow phrase searching and most are improving their Boolean searching. Few allow truncation, wild cards or nested logic. They are stateless, losing previous search criteria. Unlike the indexing and classification systems for today's CD-ROMs, those for Web pages are random, unstructured and of variable quality. Considers that at best Web search engines can only offer free text searching. Discusses whether automatic data classification systems such as Infoseek Ultra can overcome the haphazard nature of the Web with neural network technology, and whether Boolean search techniques may be redundant when replaced by technology such as the Euroferret search engine. However, artificial intelligence is rarely successful on huge, varied databases. Relevance ranking and automatic query expansion still use the same simple inverted indexes. Most Web search engines do nothing more than word counting. Further complications arise with foreign languages
  19. Slavic, A.: Interface to classification : some objectives and options (2006) 0.01
    0.0080434345 = product of:
      0.024130303 = sum of:
        0.024130303 = product of:
          0.048260607 = sum of:
            0.048260607 = weight(_text_:indexing in 2131) [ClassicSimilarity], result of:
              0.048260607 = score(doc=2131,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.2537542 = fieldWeight in 2131, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2131)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    This is a preprint to be published in the Extensions & Corrections to the UDC. The paper explains the basic functions of browsing and searching that need to be supported in relation to analytico-synthetic classifications such as Universal Decimal Classification (UDC), irrespective of any specific, real-life implementation. UDC is an example of a semi-faceted system that can be used, for instance, for both post-coordinate searching and hierarchical/facet browsing. The advantages of using a classification for IR, however, depend on the strength of the GUI, which should provide a user-friendly interface to classification browsing and searching. The power of this interface is in supporting visualisation that will 'convert' what is potentially a user-unfriendly indexing language based on symbols, to a subject presentation that is easy to understand, search and navigate. A summary of the basic functions of searching and browsing a classification that may be provided on a user-friendly interface is given and examples of classification browsing interfaces are provided.
  20. Ellis, D.; Vasconcelos, A.: ¬The relevance of facet analysis for World Wide Web subject organization and searching (2000) 0.01
    0.0080434345 = product of:
      0.024130303 = sum of:
        0.024130303 = product of:
          0.048260607 = sum of:
            0.048260607 = weight(_text_:indexing in 2477) [ClassicSimilarity], result of:
              0.048260607 = score(doc=2477,freq=2.0), product of:
                0.19018644 = queryWeight, product of:
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.049684696 = queryNorm
                0.2537542 = fieldWeight in 2477, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.8278677 = idf(docFreq=2614, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2477)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Different forms of indexing and search facilities available on the Web are described. Use of facet analysis to structure hypertext concept structures is outlined in relation to work on (1) development of hypertext knowledge bases for designers of learning materials and (2) construction of knowledge based hypertext interfaces. The problem of lack of closeness between page designers and potential users is examined. Facet analysis is suggested as a way of alleviating some difficulties associated with this problem of designing for the unknown user.

Years

Languages

Types

  • a 53
  • el 7
  • m 3
  • s 3
  • d 1
  • p 1
  • x 1
  • More… Less…

Classifications