Search (11 results, page 1 of 1)

  • × theme_ss:"Referieren"
  1. Parekh, R.L.: Advanced indexing and abstracting practices (2000) 0.07
    0.07070904 = product of:
      0.1767726 = sum of:
        0.11275144 = weight(_text_:readable in 119) [ClassicSimilarity], result of:
          0.11275144 = score(doc=119,freq=2.0), product of:
            0.2768342 = queryWeight, product of:
              6.1439276 = idf(docFreq=257, maxDocs=44218)
              0.04505818 = queryNorm
            0.4072887 = fieldWeight in 119, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.1439276 = idf(docFreq=257, maxDocs=44218)
              0.046875 = fieldNorm(doc=119)
        0.06402116 = weight(_text_:bibliographic in 119) [ClassicSimilarity], result of:
          0.06402116 = score(doc=119,freq=4.0), product of:
            0.17541347 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.04505818 = queryNorm
            0.3649729 = fieldWeight in 119, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.046875 = fieldNorm(doc=119)
      0.4 = coord(2/5)
    
    Abstract
    Indexing and abstracting are not activities that should be looked upon as ends in themselves. It is the results of these activities that should be evaluated and this can only be done within the context of a particular database, whether in printed or machine-readable form. In this context, the indexing can be judged successful if it allows searchers to locate items they want without having to look at many they do not want. This book intended primarily as a text to be used in teaching indexing and abstracting of Library and information science. It is an immense value to all individuals and institutions involved in information retrieval and related activities, including librarians, managers of information centres and database producers.
    Content
    Inhalt: 1. Indexing and Abstracting 2. Automatic Indexing and Automatic Abstracting 3. Principles of Indexing 4.Periodicals Listing and Accessioning 5. Online Computer Service 6. Dialog, Searching and Bibliographic Display 7. Books 8. Bibliographic Control 9. Abstracting Functions 10. Acquisition System 11. Future of Indexing and Abstracting Services
  2. Kuhlen, R.: Informationsaufbereitung III : Referieren (Abstracts - Abstracting - Grundlagen) (2004) 0.02
    0.015033525 = product of:
      0.075167626 = sum of:
        0.075167626 = weight(_text_:readable in 2917) [ClassicSimilarity], result of:
          0.075167626 = score(doc=2917,freq=2.0), product of:
            0.2768342 = queryWeight, product of:
              6.1439276 = idf(docFreq=257, maxDocs=44218)
              0.04505818 = queryNorm
            0.2715258 = fieldWeight in 2917, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.1439276 = idf(docFreq=257, maxDocs=44218)
              0.03125 = fieldNorm(doc=2917)
      0.2 = coord(1/5)
    
    Abstract
    Was ein Abstract (im Folgenden synonym mit Referat oder Kurzreferat gebraucht) ist, legt das American National Standards Institute in einer Weise fest, die sicherlich von den meisten Fachleuten akzeptiert werden kann: "An abstract is defined as an abbreviated, accurate representation of the contents of a document"; fast genauso die deutsche Norm DIN 1426: "Das Kurzreferat gibt kurz und klar den Inhalt des Dokuments wieder." Abstracts gehören zum wissenschaftlichen Alltag. Weitgehend allen Publikationen, zumindest in den naturwissenschaftlichen, technischen, informationsbezogenen oder medizinischen Bereichen, gehen Abstracts voran, "prefe-rably prepared by its author(s) for publication with it". Es gibt wohl keinen Wissenschaftler, der nicht irgendwann einmal ein Abstract geschrieben hätte. Gehört das Erstellen von Abstracts dann überhaupt zur dokumentarischen bzw informationswissenschaftlichen Methodenlehre, wenn es jeder kann? Was macht den informationellen Mehrwert aus, der durch Expertenreferate gegenüber Laienreferaten erzeugt wird? Dies ist nicht so leicht zu beantworten, zumal geeignete Bewertungsverfahren fehlen, die Qualität von Abstracts vergleichend "objektiv" zu messen. Abstracts werden in erheblichem Umfang von Informationsspezialisten erstellt, oft unter der Annahme, dass Autoren selber dafür weniger geeignet sind. Vergegenwärtigen wir uns, was wir über Abstracts und Abstracting wissen. Ein besonders gelungenes Abstract ist zuweilen klarer als der Ursprungstext selber, darf aber nicht mehr Information als dieser enthalten: "Good abstracts are highly structured, concise, and coherent, and are the result of a thorough analysis of the content of the abstracted materials. Abstracts may be more readable than the basis documents, but because of size constraints they rarely equal and never surpass the information content of the basic document". Dies ist verständlich, denn ein "Abstract" ist zunächst nichts anderes als ein Ergebnis des Vorgangs einer Abstraktion. Ohne uns zu sehr in die philosophischen Hintergründe der Abstraktion zu verlieren, besteht diese doch "in der Vernachlässigung von bestimmten Vorstellungsbzw. Begriffsinhalten, von welchen zugunsten anderer Teilinhalte abgesehen, abstrahiert' wird. Sie ist stets verbunden mit einer Fixierung von (interessierenden) Merkmalen durch die aktive Aufmerksamkeit, die unter einem bestimmten pragmatischen Gesichtspunkt als wesentlich' für einen vorgestellten bzw für einen unter einen Begriff fallenden Gegenstand (oder eine Mehrheit von Gegenständen) betrachtet werden". Abstracts reduzieren weniger Begriffsinhalte, sondern Texte bezüglich ihres proportionalen Gehaltes. Borko/ Bernier haben dies sogar quantifiziert; sie schätzen den Reduktionsfaktor auf 1:10 bis 1:12
  3. Wan, X.; Yang, J.; Xiao, J.: Incorporating cross-document relationships between sentences for single document summarizations (2006) 0.01
    0.0132987825 = product of:
      0.06649391 = sum of:
        0.06649391 = sum of:
          0.029865343 = weight(_text_:data in 2421) [ClassicSimilarity], result of:
            0.029865343 = score(doc=2421,freq=2.0), product of:
              0.14247625 = queryWeight, product of:
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.04505818 = queryNorm
              0.2096163 = fieldWeight in 2421, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.046875 = fieldNorm(doc=2421)
          0.036628567 = weight(_text_:22 in 2421) [ClassicSimilarity], result of:
            0.036628567 = score(doc=2421,freq=2.0), product of:
              0.15778607 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04505818 = queryNorm
              0.23214069 = fieldWeight in 2421, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2421)
      0.2 = coord(1/5)
    
    Abstract
    Graph-based ranking algorithms have recently been proposed for single document summarizations and such algorithms evaluate the importance of a sentence by making use of the relationships between sentences in the document in a recursive way. In this paper, we investigate using other related or relevant documents to improve summarization of one single document based on the graph-based ranking algorithm. In addition to the within-document relationships between sentences in the specified document, the cross-document relationships between sentences in different documents are also taken into account in the proposed approach. We evaluate the performance of the proposed approach on DUC 2002 data with the ROUGE metric and results demonstrate that the cross-document relationships between sentences in different but related documents can significantly improve the performance of single document summarization.
    Source
    Research and advanced technology for digital libraries : 10th European conference, proceedings / ECDL 2006, Alicante, Spain, September 17 - 22, 2006
  4. Hartley, J.; Sydes, M.; Blurton, A.: Obtaining information accurately and quickly : are structured abstracts more efficient? (1996) 0.01
    0.01108232 = product of:
      0.055411596 = sum of:
        0.055411596 = sum of:
          0.024887787 = weight(_text_:data in 7673) [ClassicSimilarity], result of:
            0.024887787 = score(doc=7673,freq=2.0), product of:
              0.14247625 = queryWeight, product of:
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.04505818 = queryNorm
              0.17468026 = fieldWeight in 7673, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.1620505 = idf(docFreq=5088, maxDocs=44218)
                0.0390625 = fieldNorm(doc=7673)
          0.030523809 = weight(_text_:22 in 7673) [ClassicSimilarity], result of:
            0.030523809 = score(doc=7673,freq=2.0), product of:
              0.15778607 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.04505818 = queryNorm
              0.19345059 = fieldWeight in 7673, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=7673)
      0.2 = coord(1/5)
    
    Abstract
    Reports results of 2 studies to determine if structured abstracts offer any advantage to users in terms of whether they are easier to search. In study 1, using a specially prepared electronic database of abstracts in either their original format or the structured format, 52 users were asked to find the answers to 2 questions for each of 8 abstracts in traditional format followed by 2 questions for each of 8 abstracts set in the structured format. Time and error data were recorded automatically. In study 2, using a printed database, 56 users were asked to to find 5 abstracts that reprted a particular kind of study and then find 5 more references that reported another kind of study. In study 1 users performed significantly faster and made fewer errors with structured abstracts but there were some unexplainable practice effects. In study 2, the users again performed significantly faster and made fewer errors with structured abstracts. However, there were asymmetrical transfer effects: users who responded first to the structured abstracts responded more quickly to the following traditional abstracts than did those users who responded first to the traditional abstracts. Nevertheless, the overall findings support the hypothesis that it is easier for user to search structured abstracts than it is to search traditional abstracts
    Source
    Journal of information science. 22(1996) no.5, S.349-356
  5. Montesi, M.; Mackenzie Owen, J.: Revision of author abstracts : how it is carried out by LISA editors (2007) 0.01
    0.007544966 = product of:
      0.03772483 = sum of:
        0.03772483 = weight(_text_:bibliographic in 807) [ClassicSimilarity], result of:
          0.03772483 = score(doc=807,freq=2.0), product of:
            0.17541347 = queryWeight, product of:
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.04505818 = queryNorm
            0.21506234 = fieldWeight in 807, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.893044 = idf(docFreq=2449, maxDocs=44218)
              0.0390625 = fieldNorm(doc=807)
      0.2 = coord(1/5)
    
    Abstract
    Purpose - The literature on abstracts recommends the revision of author supplied abstracts before their inclusion in database collections. However, little guidance is given on how to carry out such revision, and few studies exist on this topic. The purpose of this research paper is to first survey 187 bibliographic databases to ascertain how many did revise abstracts, and then study the practical amendments made by one of these, i.e. LISA (Library and Information Science Abstracts). Design/methodology/approach - Database policies were established by e-mail or through alternative sources, with 136 databases out of 187 exhaustively documented. Differences between 100 author-supplied abstracts and the corresponding 100 LISA amended abstracts were classified into sentence-level and beyond sentence-level categories, and then as additions, deletions and rephrasing of text. Findings - Revision of author abstracts was carried out by 66 databases, but in just 32 cases did it imply more than spelling, shortening of length and formula representation. In LISA, amendments were often non-systematic and inconsistent, but still pointed to significant aspects which were discussed. Originality/value - Amendments made by LISA editors are important in multi- and inter-disciplinary research, since they tend to clarify certain aspects such as terminology, and suggest that abstracts should not always be considered as substitutes for the original document. From this point-of-view, the revision of abstracts can be considered as an important factor in enhancing a database's quality.
  6. Koltay, T.: ¬A hypertext tutorial on abstracting for library science students (1995) 0.01
    0.0061047617 = product of:
      0.030523809 = sum of:
        0.030523809 = product of:
          0.061047617 = sum of:
            0.061047617 = weight(_text_:22 in 3061) [ClassicSimilarity], result of:
              0.061047617 = score(doc=3061,freq=2.0), product of:
                0.15778607 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04505818 = queryNorm
                0.38690117 = fieldWeight in 3061, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3061)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Date
    27. 1.1996 18:22:06
  7. Palais, E.S.: Abstracting for reference librarians (1988) 0.00
    0.004883809 = product of:
      0.024419045 = sum of:
        0.024419045 = product of:
          0.04883809 = sum of:
            0.04883809 = weight(_text_:22 in 2832) [ClassicSimilarity], result of:
              0.04883809 = score(doc=2832,freq=2.0), product of:
                0.15778607 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04505818 = queryNorm
                0.30952093 = fieldWeight in 2832, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2832)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Source
    Reference librarian. 1988, no.22, S.297-308
  8. Molina, M.P.: Documentary abstracting : toward a methodological approach (1995) 0.00
    0.003982046 = product of:
      0.01991023 = sum of:
        0.01991023 = product of:
          0.03982046 = sum of:
            0.03982046 = weight(_text_:data in 1790) [ClassicSimilarity], result of:
              0.03982046 = score(doc=1790,freq=2.0), product of:
                0.14247625 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.04505818 = queryNorm
                0.2794884 = fieldWeight in 1790, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1790)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Abstract
    In the general abstracting process (GAP), there are 2 types of data: textual, within a particular framed trilogy (surface, deep, and rhetoric); and documentary (abstractor, means of production, and user demands). Proposes its development, the use of the following disciplines, among others: linguistics (structural, tranformational, and textual), logic (formal and fuzzy), and psychology (cognitive). The model for that textual transformation is based on a system of combined strategies with 4 key stages: reading understanding, selection, interpretation, and synthesis
  9. Hartley, J.; Sydes, M.: Which layout do you prefer? : an analysis of readers' preferences for different typographic layouts of structured abstracts (1996) 0.00
    0.0036628568 = product of:
      0.018314283 = sum of:
        0.018314283 = product of:
          0.036628567 = sum of:
            0.036628567 = weight(_text_:22 in 4411) [ClassicSimilarity], result of:
              0.036628567 = score(doc=4411,freq=2.0), product of:
                0.15778607 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04505818 = queryNorm
                0.23214069 = fieldWeight in 4411, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4411)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Source
    Journal of information science. 22(1996) no.1, S.27-37
  10. Ward, M.L.: ¬The future of the human indexer (1996) 0.00
    0.0036628568 = product of:
      0.018314283 = sum of:
        0.018314283 = product of:
          0.036628567 = sum of:
            0.036628567 = weight(_text_:22 in 7244) [ClassicSimilarity], result of:
              0.036628567 = score(doc=7244,freq=2.0), product of:
                0.15778607 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.04505818 = queryNorm
                0.23214069 = fieldWeight in 7244, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=7244)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Date
    9. 2.1997 18:44:22
  11. Hartley, J.; Betts, L.: Revising and polishing a structured abstract : is it worth the time and effort? (2008) 0.00
    0.0024887787 = product of:
      0.012443894 = sum of:
        0.012443894 = product of:
          0.024887787 = sum of:
            0.024887787 = weight(_text_:data in 2362) [ClassicSimilarity], result of:
              0.024887787 = score(doc=2362,freq=2.0), product of:
                0.14247625 = queryWeight, product of:
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.04505818 = queryNorm
                0.17468026 = fieldWeight in 2362, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.1620505 = idf(docFreq=5088, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2362)
          0.5 = coord(1/2)
      0.2 = coord(1/5)
    
    Abstract
    Many writers of structured abstracts spend a good deal of time revising and polishing their texts - but is it worth it? Do readers notice the difference? In this paper we report three studies of readers using rating scales to judge (electronically) the clarity of an original and a revised abstract, both as a whole and in its constituent parts. In Study 1, with approximately 250 academics and research workers, we found some significant differences in favor of the revised abstract, but in Study 2, with approximately 210 information scientists, we found no significant effects. Pooling the data from Studies 1 and 2, however, in Study 3, led to significant differences at a higher probability level between the perception of the original and revised abstract as a whole and between the same components as found in Study 1. These results thus indicate that the revised abstract as a whole, as well as certain specific components of it, were judged significantly clearer than the original one. In short, the results of these experiments show that readers can and do perceive differences between original and revised texts - sometimes - and that therefore these efforts are worth the time and effort.