Search (80 results, page 1 of 4)

  • × type_ss:"a"
  • × type_ss:"el"
  • × year_i:[2000 TO 2010}
  1. Bittner, T.; Donnelly, M.; Winter, S.: Ontology and semantic interoperability (2006) 0.04
    0.035469428 = product of:
      0.14187771 = sum of:
        0.036391225 = weight(_text_:software in 4820) [ClassicSimilarity], result of:
          0.036391225 = score(doc=4820,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.4555077 = fieldWeight in 4820, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=4820)
        0.036391225 = weight(_text_:software in 4820) [ClassicSimilarity], result of:
          0.036391225 = score(doc=4820,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.4555077 = fieldWeight in 4820, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=4820)
        0.036391225 = weight(_text_:software in 4820) [ClassicSimilarity], result of:
          0.036391225 = score(doc=4820,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.4555077 = fieldWeight in 4820, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=4820)
        0.027247135 = product of:
          0.05449427 = sum of:
            0.05449427 = weight(_text_:engineering in 4820) [ClassicSimilarity], result of:
              0.05449427 = score(doc=4820,freq=4.0), product of:
                0.10819342 = queryWeight, product of:
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.02013827 = queryNorm
                0.5036745 = fieldWeight in 4820, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4820)
          0.5 = coord(1/2)
        0.005456915 = product of:
          0.016370745 = sum of:
            0.016370745 = weight(_text_:22 in 4820) [ClassicSimilarity], result of:
              0.016370745 = score(doc=4820,freq=2.0), product of:
                0.07052079 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02013827 = queryNorm
                0.23214069 = fieldWeight in 4820, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4820)
          0.33333334 = coord(1/3)
      0.25 = coord(5/20)
    
    Abstract
    One of the major problems facing systems for Computer Aided Design (CAD), Architecture Engineering and Construction (AEC) and Geographic Information Systems (GIS) applications today is the lack of interoperability among the various systems. When integrating software applications, substantial di culties can arise in translating information from one application to the other. In this paper, we focus on semantic di culties that arise in software integration. Applications may use di erent terminologies to describe the same domain. Even when appli-cations use the same terminology, they often associate di erent semantics with the terms. This obstructs information exchange among applications. To cir-cumvent this obstacle, we need some way of explicitly specifying the semantics for each terminology in an unambiguous fashion. Ontologies can provide such specification. It will be the task of this paper to explain what ontologies are and how they can be used to facilitate interoperability between software systems used in computer aided design, architecture engineering and construction, and geographic information processing.
    Date
    3.12.2016 18:39:22
  2. Hoffmann, P.; Médini and , L.; Ghodous, P.: Using context to improve semantic interoperability (2006) 0.02
    0.024762774 = product of:
      0.12381387 = sum of:
        0.02829376 = weight(_text_:23 in 4434) [ClassicSimilarity], result of:
          0.02829376 = score(doc=4434,freq=4.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.39200652 = fieldWeight in 4434, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4434)
        0.02829376 = weight(_text_:23 in 4434) [ClassicSimilarity], result of:
          0.02829376 = score(doc=4434,freq=4.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.39200652 = fieldWeight in 4434, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4434)
        0.02829376 = weight(_text_:23 in 4434) [ClassicSimilarity], result of:
          0.02829376 = score(doc=4434,freq=4.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.39200652 = fieldWeight in 4434, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4434)
        0.038932588 = product of:
          0.077865176 = sum of:
            0.077865176 = weight(_text_:engineering in 4434) [ClassicSimilarity], result of:
              0.077865176 = score(doc=4434,freq=6.0), product of:
                0.10819342 = queryWeight, product of:
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.02013827 = queryNorm
                0.71968496 = fieldWeight in 4434, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4434)
          0.5 = coord(1/2)
      0.2 = coord(4/20)
    
    Date
    23. 6.2011 13:23:58
    Source
    Leading the Web in concurrent engineering: next generation concurrent engineering. Proceeding of the 2006 ISPE Conference on Concurrent Engineering. Edited by Parisa Ghodous, Rose Dieng-Kuntz, Geilson Loureiro
  3. Witten, I.H.; Bainbridge, D.; Boddie, S.J.: Greenstone : open-source digital library software (2001) 0.02
    0.016376052 = product of:
      0.10917367 = sum of:
        0.036391225 = weight(_text_:software in 1225) [ClassicSimilarity], result of:
          0.036391225 = score(doc=1225,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.4555077 = fieldWeight in 1225, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=1225)
        0.036391225 = weight(_text_:software in 1225) [ClassicSimilarity], result of:
          0.036391225 = score(doc=1225,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.4555077 = fieldWeight in 1225, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=1225)
        0.036391225 = weight(_text_:software in 1225) [ClassicSimilarity], result of:
          0.036391225 = score(doc=1225,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.4555077 = fieldWeight in 1225, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=1225)
      0.15 = coord(3/20)
    
    Abstract
    The Greenstone digital library software is an open-source system for the construction and presentation of information collections. It builds collections with effective full-text searching and metadata-based browsing facilities that are attractive and easy to use. Moreover, they are easily maintained and can be augmented and rebuilt entirely automatically. The system is extensible: software "plugins" accommodate different document and metadata types. Greenstone incorporates an interface that makes it easy for people to create their own library collections. Collections may be built and served locally from the user's own web server, or (given appropriate permissions) remotely on a shared digital library host. End users can easily build new collections styled after existing ones from material on the Web or from their local files (or both), and collections can be updated and new ones brought on-line at any time.
  4. Atran, S.; Medin, D.L.; Ross, N.: Evolution and devolution of knowledge : a tale of two biologies (2004) 0.02
    0.015642462 = product of:
      0.078212306 = sum of:
        0.024251796 = weight(_text_:23 in 479) [ClassicSimilarity], result of:
          0.024251796 = score(doc=479,freq=4.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.3360056 = fieldWeight in 479, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=479)
        0.024251796 = weight(_text_:23 in 479) [ClassicSimilarity], result of:
          0.024251796 = score(doc=479,freq=4.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.3360056 = fieldWeight in 479, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=479)
        0.024251796 = weight(_text_:23 in 479) [ClassicSimilarity], result of:
          0.024251796 = score(doc=479,freq=4.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.3360056 = fieldWeight in 479, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=479)
        0.005456915 = product of:
          0.016370745 = sum of:
            0.016370745 = weight(_text_:22 in 479) [ClassicSimilarity], result of:
              0.016370745 = score(doc=479,freq=2.0), product of:
                0.07052079 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02013827 = queryNorm
                0.23214069 = fieldWeight in 479, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.046875 = fieldNorm(doc=479)
          0.33333334 = coord(1/3)
      0.2 = coord(4/20)
    
    Date
    23. 1.2022 10:08:57
    23. 1.2022 10:22:18
  5. Mongin, L.; Fu, Y.Y.; Mostafa, J.: Open Archives data Service prototype and automated subject indexing using D-Lib archive content as a testbed (2003) 0.01
    0.013370992 = product of:
      0.08913994 = sum of:
        0.029713312 = weight(_text_:software in 1167) [ClassicSimilarity], result of:
          0.029713312 = score(doc=1167,freq=4.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.3719205 = fieldWeight in 1167, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=1167)
        0.029713312 = weight(_text_:software in 1167) [ClassicSimilarity], result of:
          0.029713312 = score(doc=1167,freq=4.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.3719205 = fieldWeight in 1167, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=1167)
        0.029713312 = weight(_text_:software in 1167) [ClassicSimilarity], result of:
          0.029713312 = score(doc=1167,freq=4.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.3719205 = fieldWeight in 1167, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=1167)
      0.15 = coord(3/20)
    
    Abstract
    The Indiana University School of Library and Information Science opened a new research laboratory in January 2003; The Indiana University School of Library and Information Science Information Processing Laboratory [IU IP Lab]. The purpose of the new laboratory is to facilitate collaboration between scientists in the department in the areas of information retrieval (IR) and information visualization (IV) research. The lab has several areas of focus. These include grid and cluster computing, and a standard Java-based software platform to support plug and play research datasets, a selection of standard IR modules and standard IV algorithms. Future development includes software to enable researchers to contribute datasets, IR algorithms, and visualization algorithms into the standard environment. We decided early on to use OAI-PMH as a resource discovery tool because it is consistent with our mission.
  6. Reiner, U.: Automatische DDC-Klassifizierung bibliografischer Titeldatensätze der Deutschen Nationalbibliografie (2009) 0.01
    0.010972724 = product of:
      0.05486362 = sum of:
        0.010708931 = weight(_text_:und in 3284) [ClassicSimilarity], result of:
          0.010708931 = score(doc=3284,freq=12.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.23992877 = fieldWeight in 3284, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.03125 = fieldNorm(doc=3284)
        0.023900734 = weight(_text_:methoden in 3284) [ClassicSimilarity], result of:
          0.023900734 = score(doc=3284,freq=2.0), product of:
            0.10436003 = queryWeight, product of:
              5.1821747 = idf(docFreq=674, maxDocs=44218)
              0.02013827 = queryNorm
            0.22902192 = fieldWeight in 3284, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.1821747 = idf(docFreq=674, maxDocs=44218)
              0.03125 = fieldNorm(doc=3284)
        0.016616013 = weight(_text_:der in 3284) [ClassicSimilarity], result of:
          0.016616013 = score(doc=3284,freq=28.0), product of:
            0.044984195 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.02013827 = queryNorm
            0.36937445 = fieldWeight in 3284, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.03125 = fieldNorm(doc=3284)
        0.0036379434 = product of:
          0.01091383 = sum of:
            0.01091383 = weight(_text_:22 in 3284) [ClassicSimilarity], result of:
              0.01091383 = score(doc=3284,freq=2.0), product of:
                0.07052079 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02013827 = queryNorm
                0.15476047 = fieldWeight in 3284, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=3284)
          0.33333334 = coord(1/3)
      0.2 = coord(4/20)
    
    Abstract
    Das Klassifizieren von Objekten (z. B. Fauna, Flora, Texte) ist ein Verfahren, das auf menschlicher Intelligenz basiert. In der Informatik - insbesondere im Gebiet der Künstlichen Intelligenz (KI) - wird u. a. untersucht, inweit Verfahren, die menschliche Intelligenz benötigen, automatisiert werden können. Hierbei hat sich herausgestellt, dass die Lösung von Alltagsproblemen eine größere Herausforderung darstellt, als die Lösung von Spezialproblemen, wie z. B. das Erstellen eines Schachcomputers. So ist "Rybka" der seit Juni 2007 amtierende Computerschach-Weltmeistern. Inwieweit Alltagsprobleme mit Methoden der Künstlichen Intelligenz gelöst werden können, ist eine - für den allgemeinen Fall - noch offene Frage. Beim Lösen von Alltagsproblemen spielt die Verarbeitung der natürlichen Sprache, wie z. B. das Verstehen, eine wesentliche Rolle. Den "gesunden Menschenverstand" als Maschine (in der Cyc-Wissensbasis in Form von Fakten und Regeln) zu realisieren, ist Lenat's Ziel seit 1984. Bezüglich des KI-Paradeprojektes "Cyc" gibt es CycOptimisten und Cyc-Pessimisten. Das Verstehen der natürlichen Sprache (z. B. Werktitel, Zusammenfassung, Vorwort, Inhalt) ist auch beim intellektuellen Klassifizieren von bibliografischen Titeldatensätzen oder Netzpublikationen notwendig, um diese Textobjekte korrekt klassifizieren zu können. Seit dem Jahr 2007 werden von der Deutschen Nationalbibliothek nahezu alle Veröffentlichungen mit der Dewey Dezimalklassifikation (DDC) intellektuell klassifiziert.
    Die Menge der zu klassifizierenden Veröffentlichungen steigt spätestens seit der Existenz des World Wide Web schneller an, als sie intellektuell sachlich erschlossen werden kann. Daher werden Verfahren gesucht, um die Klassifizierung von Textobjekten zu automatisieren oder die intellektuelle Klassifizierung zumindest zu unterstützen. Seit 1968 gibt es Verfahren zur automatischen Dokumentenklassifizierung (Information Retrieval, kurz: IR) und seit 1992 zur automatischen Textklassifizierung (ATC: Automated Text Categorization). Seit immer mehr digitale Objekte im World Wide Web zur Verfügung stehen, haben Arbeiten zur automatischen Textklassifizierung seit ca. 1998 verstärkt zugenommen. Dazu gehören seit 1996 auch Arbeiten zur automatischen DDC-Klassifizierung bzw. RVK-Klassifizierung von bibliografischen Titeldatensätzen und Volltextdokumenten. Bei den Entwicklungen handelt es sich unseres Wissens bislang um experimentelle und keine im ständigen Betrieb befindlichen Systeme. Auch das VZG-Projekt Colibri/DDC ist seit 2006 u. a. mit der automatischen DDC-Klassifizierung befasst. Die diesbezüglichen Untersuchungen und Entwicklungen dienen zur Beantwortung der Forschungsfrage: "Ist es möglich, eine inhaltlich stimmige DDC-Titelklassifikation aller GVK-PLUS-Titeldatensätze automatisch zu erzielen?"
    Date
    22. 1.2010 14:41:24
  7. DiLauro, T.; Choudhury, G.S.; Patton, M.; Warner, J.W.; Brown, E.W.: Automated name authority control and enhanced searching in the Levy collection (2001) 0.01
    0.010917367 = product of:
      0.07278245 = sum of:
        0.024260817 = weight(_text_:software in 1160) [ClassicSimilarity], result of:
          0.024260817 = score(doc=1160,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.3036718 = fieldWeight in 1160, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03125 = fieldNorm(doc=1160)
        0.024260817 = weight(_text_:software in 1160) [ClassicSimilarity], result of:
          0.024260817 = score(doc=1160,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.3036718 = fieldWeight in 1160, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03125 = fieldNorm(doc=1160)
        0.024260817 = weight(_text_:software in 1160) [ClassicSimilarity], result of:
          0.024260817 = score(doc=1160,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.3036718 = fieldWeight in 1160, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03125 = fieldNorm(doc=1160)
      0.15 = coord(3/20)
    
    Abstract
    This paper is the second in a series in D-Lib Magazine and describes a workflow management system being developed by the Digital Knowledge Center (DKC) at the Milton S. Eisenhower Library (MSEL) of The Johns Hopkins University. Based on experience from digitizing the Lester S. Levy Collection of Sheet Music, it was apparent that large-scale digitization efforts require a significant amount of human labor that is both time-consuming and costly. Consequently, this workflow management system aims to reduce the amount of human labor and time for large-scale digitization projects. The mission of this second phase of the project ("Levy II") can be summarized as follows: * Reduce costs for large collection ingestion by creating a suite of open-source processes, tools, and interfaces for workflow management * Increase access capabilities by providing a suite of research tools * Demonstrate utility of tools and processes with a subset of the online Levy Collection The cornerstones of the workflow management system include optical music recognition (OMR) software and an automated name authority control system (ANAC). The OMR software generates a logical representation of the score for sound generation, music searching, and musicological research. The ANAC disambiguates names, associating each name with an individual (e.g., the composer Septimus Winner also published under the pseudonyms Alice Hawthorne and Apsley Street, among others). Complementing the workflow tools, a suite of research tools focuses upon enhanced searching capabilities through the development and application of a fast, disk-based search engine for lyrics and music and the incorporation of an XML structure for metadata. The first paper (Choudhury et al. 2001) described the OMR software and musical components of Levy II. This paper focuses on the metadata and intellectual access components that include automated name authority control and the aforementioned search engine.
  8. Crane, G.; Jones, A.: Text, information, knowledge and the evolving record of humanity (2006) 0.01
    0.01070336 = product of:
      0.053516798 = sum of:
        0.015163011 = weight(_text_:software in 1182) [ClassicSimilarity], result of:
          0.015163011 = score(doc=1182,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.18979488 = fieldWeight in 1182, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.01953125 = fieldNorm(doc=1182)
        0.015163011 = weight(_text_:software in 1182) [ClassicSimilarity], result of:
          0.015163011 = score(doc=1182,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.18979488 = fieldWeight in 1182, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.01953125 = fieldNorm(doc=1182)
        0.015163011 = weight(_text_:software in 1182) [ClassicSimilarity], result of:
          0.015163011 = score(doc=1182,freq=6.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.18979488 = fieldWeight in 1182, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.01953125 = fieldNorm(doc=1182)
        0.008027764 = product of:
          0.016055528 = sum of:
            0.016055528 = weight(_text_:engineering in 1182) [ClassicSimilarity], result of:
              0.016055528 = score(doc=1182,freq=2.0), product of:
                0.10819342 = queryWeight, product of:
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.02013827 = queryNorm
                0.14839652 = fieldWeight in 1182, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.372528 = idf(docFreq=557, maxDocs=44218)
                  0.01953125 = fieldNorm(doc=1182)
          0.5 = coord(1/2)
      0.2 = coord(4/20)
    
    Abstract
    Consider a sentence such as "the current price of tea in China is 35 cents per pound." In a library with millions of books we might find many statements of the above form that we could capture today with relatively simple rules: rather than pursuing every variation of a statement, programs can wait, like predators at a water hole, for their informational prey to reappear in a standard linguistic pattern. We can make inferences from sentences such as "NAME1 born at NAME2 in DATE" that NAME more likely than not represents a person and NAME a place and then convert the statement into a proposition about a person born at a given place and time. The changing price of tea in China, pedestrian birth and death dates, or other basic statements may not be truth and beauty in the Phaedrus, but a digital library that could plot the prices of various commodities in different markets over time, plot the various lifetimes of individuals, or extract and classify many events would be very useful. Services such as the Syllabus Finder1 and H-Bot2 (which Dan Cohen describes elsewhere in this issue of D-Lib) represent examples of information extraction already in use. H-Bot, in particular, builds on our evolving ability to extract information from very large corpora such as the billions of web pages available through the Google API. Aside from identifying higher order statements, however, users also want to search and browse named entities: they want to read about "C. P. E. Bach" rather than his father "Johann Sebastian" or about "Cambridge, Maryland", without hearing about "Cambridge, Massachusetts", Cambridge in the UK or any of the other Cambridges scattered around the world. Named entity identification is a well-established area with an ongoing literature. The Natural Language Processing Research Group at the University of Sheffield has developed its open source Generalized Architecture for Text Engineering (GATE) for years, while IBM's Unstructured Information Analysis and Search (UIMA) is "available as open source software to provide a common foundation for industry and academia." Powerful tools are thus freely available and more demanding users can draw upon published literature to develop their own systems. Major search engines such as Google and Yahoo also integrate increasingly sophisticated tools to categorize and identify places. The software resources are rich and expanding. The reference works on which these systems depend, however, are ill-suited for historical analysis. First, simple gazetteers and similar authority lists quickly grow too big for useful information extraction. They provide us with potential entities against which to match textual references, but existing electronic reference works assume that human readers can use their knowledge of geography and of the immediate context to pick the right Boston from the Bostons in the Getty Thesaurus of Geographic Names (TGN), but, with the crucial exception of geographic location, the TGN records do not provide any machine readable clues: we cannot tell which Bostons are large or small. If we are analyzing a document published in 1818, we cannot filter out those places that did not yet exist or that had different names: "Jefferson Davis" is not the name of a parish in Louisiana (tgn,2000880) or a county in Mississippi (tgn,2001118) until after the Civil War.
    Although the Alexandria Digital Library provides far richer data than the TGN (5.9 vs. 1.3 million names), its added size lowers, rather than increases, the accuracy of most geographic name identification systems for historical documents: most of the extra 4.6 million names cover low frequency entities that rarely occur in any particular corpus. The TGN is sufficiently comprehensive to provide quite enough noise: we find place names that are used over and over (there are almost one hundred Washingtons) and semantically ambiguous (e.g., is Washington a person or a place?). Comprehensive knowledge sources emphasize recall but lower precision. We need data with which to determine which "Tribune" or "John Brown" a particular passage denotes. Secondly and paradoxically, our reference works may not be comprehensive enough. Human actors come and go over time. Organizations appear and vanish. Even places can change their names or vanish. The TGN does associate the obsolete name Siam with the nation of Thailand (tgn,1000142) - but also with towns named Siam in Iowa (tgn,2035651), Tennessee (tgn,2101519), and Ohio (tgn,2662003). Prussia appears but as a general region (tgn,7016786), with no indication when or if it was a sovereign nation. And if places do point to the same object over time, that object may have very different significance over time: in the foundational works of Western historiography, Herodotus reminds us that the great cities of the past may be small today, and the small cities of today great tomorrow (Hdt. 1.5), while Thucydides stresses that we cannot estimate the past significance of a place by its appearance today (Thuc. 1.10). In other words, we need to know the population figures for the various Washingtons in 1870 if we are analyzing documents from 1870. The foundations have been laid for reference works that provide machine actionable information about entities at particular times in history. The Alexandria Digital Library Gazetteer Content Standard8 represents a sophisticated framework with which to create such resources: places can be associated with temporal information about their foundation (e.g., Washington, DC, founded on 16 July 1790), changes in names for the same location (e.g., Saint Petersburg to Leningrad and back again), population figures at various times and similar historically contingent data. But if we have the software and the data structures, we do not yet have substantial amounts of historical content such as plentiful digital gazetteers, encyclopedias, lexica, grammars and other reference works to illustrate many periods and, even if we do, those resources may not be in a useful form: raw OCR output of a complex lexicon or gazetteer may have so many errors and have captured so little of the underlying structure that the digital resource is useless as a knowledge base. Put another way, human beings are still much better at reading and interpreting the contents of page images than machines. While people, places, and dates are probably the most important core entities, we will find a growing set of objects that we need to identify and track across collections, and each of these categories of objects will require its own knowledge sources. The following section enumerates and briefly describes some existing categories of documents that we need to mine for knowledge. This brief survey focuses on the format of print sources (e.g., highly structured textual "database" vs. unstructured text) to illustrate some of the challenges involved in converting our published knowledge into semantically annotated, machine actionable form.
  9. Hoang, H.H.; Tjoa, A.M: ¬The state of the art of ontology-based query systems : a comparison of existing approaches (2006) 0.01
    0.010289166 = product of:
      0.06859444 = sum of:
        0.022864813 = weight(_text_:23 in 792) [ClassicSimilarity], result of:
          0.022864813 = score(doc=792,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.31678912 = fieldWeight in 792, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0625 = fieldNorm(doc=792)
        0.022864813 = weight(_text_:23 in 792) [ClassicSimilarity], result of:
          0.022864813 = score(doc=792,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.31678912 = fieldWeight in 792, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0625 = fieldNorm(doc=792)
        0.022864813 = weight(_text_:23 in 792) [ClassicSimilarity], result of:
          0.022864813 = score(doc=792,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.31678912 = fieldWeight in 792, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0625 = fieldNorm(doc=792)
      0.15 = coord(3/20)
    
    Date
    28.11.2016 18:23:55
  10. Shechtman, N.; Chung, M.; Roschelle, J.: Supporting member collaboration in the Math Tools digital library : a formative user study (2004) 0.01
    0.009454718 = product of:
      0.06303145 = sum of:
        0.021010485 = weight(_text_:software in 1163) [ClassicSimilarity], result of:
          0.021010485 = score(doc=1163,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.2629875 = fieldWeight in 1163, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=1163)
        0.021010485 = weight(_text_:software in 1163) [ClassicSimilarity], result of:
          0.021010485 = score(doc=1163,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.2629875 = fieldWeight in 1163, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=1163)
        0.021010485 = weight(_text_:software in 1163) [ClassicSimilarity], result of:
          0.021010485 = score(doc=1163,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.2629875 = fieldWeight in 1163, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=1163)
      0.15 = coord(3/20)
    
    Abstract
    In this paper, we discuss a user study done at the formative stage of development of a Math Tools developers' community. The Math Tools digital library, which aims to collect software tools to support K-12 and university mathematics instruction, has two synergistic purposes. One is to support federated search and the other is to create a community of practice in which developers and users can work together. While much research has explored the technical problem of federated search, there has been little investigation into how to grow a creative, working community around a digital library. To this end, we surveyed and interviewed members of the Math Tools community in order to elicit concerns and priorities. These data led to rich descriptions of the teachers, developers, and researchers who comprise this community. Insights from these descriptions were then used to inform the creation of a set of metaphors and design principles that the Math Tools team could use in their continuing design work.
  11. Bradford, R.B.: Relationship discovery in large text collections using Latent Semantic Indexing (2006) 0.01
    0.009131783 = product of:
      0.045658913 = sum of:
        0.014006989 = weight(_text_:software in 1163) [ClassicSimilarity], result of:
          0.014006989 = score(doc=1163,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.17532499 = fieldWeight in 1163, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03125 = fieldNorm(doc=1163)
        0.014006989 = weight(_text_:software in 1163) [ClassicSimilarity], result of:
          0.014006989 = score(doc=1163,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.17532499 = fieldWeight in 1163, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03125 = fieldNorm(doc=1163)
        0.014006989 = weight(_text_:software in 1163) [ClassicSimilarity], result of:
          0.014006989 = score(doc=1163,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.17532499 = fieldWeight in 1163, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03125 = fieldNorm(doc=1163)
        0.0036379434 = product of:
          0.01091383 = sum of:
            0.01091383 = weight(_text_:22 in 1163) [ClassicSimilarity], result of:
              0.01091383 = score(doc=1163,freq=2.0), product of:
                0.07052079 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02013827 = queryNorm
                0.15476047 = fieldWeight in 1163, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1163)
          0.33333334 = coord(1/3)
      0.2 = coord(4/20)
    
    Abstract
    This paper addresses the problem of information discovery in large collections of text. For users, one of the key problems in working with such collections is determining where to focus their attention. In selecting documents for examination, users must be able to formulate reasonably precise queries. Queries that are too broad will greatly reduce the efficiency of information discovery efforts by overwhelming the users with peripheral information. In order to formulate efficient queries, a mechanism is needed to automatically alert users regarding potentially interesting information contained within the collection. This paper presents the results of an experiment designed to test one approach to generation of such alerts. The technique of latent semantic indexing (LSI) is used to identify relationships among entities of interest. Entity extraction software is used to pre-process the text of the collection so that the LSI space contains representation vectors for named entities in addition to those for individual terms. In the LSI space, the cosine of the angle between the representation vectors for two entities captures important information regarding the degree of association of those two entities. For appropriate choices of entities, determining the entity pairs with the highest mutual cosine values yields valuable information regarding the contents of the text collection. The test database used for the experiment consists of 150,000 news articles. The proposed approach for alert generation is tested using a counterterrorism analysis example. The approach is shown to have significant potential for aiding users in rapidly focusing on information of potential importance in large text collections. The approach also has value in identifying possible use of aliases.
    Source
    Proceedings of the Fourth Workshop on Link Analysis, Counterterrorism, and Security, SIAM Data Mining Conference, Bethesda, MD, 20-22 April, 2006. [http://www.siam.org/meetings/sdm06/workproceed/Link%20Analysis/15.pdf]
  12. Schaefer, A.; Jordan, M.; Klas, C.-P.; Fuhr, N.: Active support for query formulation in virtual digital libraries : a case study with DAFFODIL (2005) 0.01
    0.009094424 = product of:
      0.06062949 = sum of:
        0.02020983 = weight(_text_:23 in 4296) [ClassicSimilarity], result of:
          0.02020983 = score(doc=4296,freq=4.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.28000468 = fieldWeight in 4296, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4296)
        0.02020983 = weight(_text_:23 in 4296) [ClassicSimilarity], result of:
          0.02020983 = score(doc=4296,freq=4.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.28000468 = fieldWeight in 4296, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4296)
        0.02020983 = weight(_text_:23 in 4296) [ClassicSimilarity], result of:
          0.02020983 = score(doc=4296,freq=4.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.28000468 = fieldWeight in 4296, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4296)
      0.15 = coord(3/20)
    
    Date
    16.11.2008 17:23:50
    Source
    Research and advanced technology for digital libraries : 9th European conference, ECDL 2005, Vienna, Austria, September 18-23, 2005 ; proceedings / Andreas Rauber ... (eds.)
  13. Buckland, M.; Lancaster, L.: Combining place, time, and topic : the Electronic Cultural Atlas Initiative (2004) 0.01
    0.008913995 = product of:
      0.059426628 = sum of:
        0.019808875 = weight(_text_:software in 1194) [ClassicSimilarity], result of:
          0.019808875 = score(doc=1194,freq=4.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.24794699 = fieldWeight in 1194, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03125 = fieldNorm(doc=1194)
        0.019808875 = weight(_text_:software in 1194) [ClassicSimilarity], result of:
          0.019808875 = score(doc=1194,freq=4.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.24794699 = fieldWeight in 1194, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03125 = fieldNorm(doc=1194)
        0.019808875 = weight(_text_:software in 1194) [ClassicSimilarity], result of:
          0.019808875 = score(doc=1194,freq=4.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.24794699 = fieldWeight in 1194, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03125 = fieldNorm(doc=1194)
      0.15 = coord(3/20)
    
    Abstract
    The Electronic Cultural Atlas Initiative was formed to encourage scholarly communication and the sharing of data among researchers who emphasize the relationships between place, time, and topic in the study of culture and history. In an effort to develop better tools and practices, The Electronic Cultural Atlas Initiative has sponsored the collaborative development of software for downloading and editing geo-temporal data to create dynamic maps, a clearinghouse of shared datasets accessible through a map-based interface, projects on format and content standards for gazetteers and time period directories, studies to improve geo-temporal aspects in online catalogs, good practice guidelines for preparing e-publications with dynamic geo-temporal displays, and numerous international conferences. The Electronic Cultural Atlas Initiative (ECAI) grew out of discussions among an international group of scholars interested in religious history and area studies. It was established as a unit under the Dean of International and Area Studies at the University of California, Berkeley in 1997. ECAI's mission is to promote an international collaborative effort to transform humanities scholarship through use of the digital environment to share data and by placing greater emphasis on the notions of place and time. Professor Lewis Lancaster is the Director. Professor Michael Buckland, with a library and information studies background, joined the effort as Co-Director in 2000. Assistance from the Lilly Foundation, the California Digital Library (University of California), and other sources has enabled ECAI to nurture a community; to develop a catalog ("clearinghouse") of Internet-accessible georeferenced resources; to support the development of software for obtaining, editing, manipulating, and dynamically visualizing geo-temporally encoded data; and to undertake research and development projects as needs and resources determine. Several hundred scholars worldwide, from a wide range of disciplines, are informally affiliated with ECAI, all interested in shared use of historical and cultural data. The Academia Sinica (Taiwan), The British Library, and the Arts and Humanities Data Service (UK) are among the well-known affiliates. However, ECAI mainly comprises individual scholars and small teams working on their own small projects on a very wide range of cultural, social, and historical topics. Numerous specialist committees have been fostering standardization and collaboration by area and by themes such as trade-routes, cities, religion, and sacred sites.
  14. Eversberg, B.: Zur Zukunft der Katalogisierung : ... jenseits RAK und AACR (2004) 0.01
    0.008490006 = product of:
      0.05660004 = sum of:
        0.021639816 = weight(_text_:und in 3632) [ClassicSimilarity], result of:
          0.021639816 = score(doc=3632,freq=16.0), product of:
            0.044633795 = queryWeight, product of:
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.02013827 = queryNorm
            0.4848303 = fieldWeight in 3632, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              2.216367 = idf(docFreq=13101, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3632)
        0.021499714 = product of:
          0.042999428 = sum of:
            0.042999428 = weight(_text_:allgemein in 3632) [ClassicSimilarity], result of:
              0.042999428 = score(doc=3632,freq=2.0), product of:
                0.10581345 = queryWeight, product of:
                  5.254347 = idf(docFreq=627, maxDocs=44218)
                  0.02013827 = queryNorm
                0.40637016 = fieldWeight in 3632, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  5.254347 = idf(docFreq=627, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=3632)
          0.5 = coord(1/2)
        0.0134605095 = weight(_text_:der in 3632) [ClassicSimilarity], result of:
          0.0134605095 = score(doc=3632,freq=6.0), product of:
            0.044984195 = queryWeight, product of:
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.02013827 = queryNorm
            0.29922754 = fieldWeight in 3632, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              2.2337668 = idf(docFreq=12875, maxDocs=44218)
              0.0546875 = fieldNorm(doc=3632)
      0.15 = coord(3/20)
    
    Abstract
    "Katalogisierung" klingt in manchen Ohren altmodisch. Man redet heute von "Metadaten"! Gemeint ist aber nichts völlig anderes. Es gibt nur viele neue Formen, Praktiken und Nutzungen, während sich früher Katalogdaten fast nur in Katalogen und Bibliographien befanden. "Metadaten" ist nur ein neuer Oberbegriff, aber in einer Katalogdatenbank haben wir längst mehr und andersartige Datenelemente und Funktionen als in Zettelkatalogen. Es ist notwendig, weiter auszugreifen als es die klassischen Regelwerke, RAK und AACR, getan haben, und deren hergebrachte Konzepte zu überdenken.
    Footnote
    Präsentation zum Vortrag "Zur Zukunft der Katalogisierung" während des Österreichischen Bibliothekartages in Linz 22.09.2004, Themenkreis: Google und die Zukunft der bibliothekarischen Erschließung. - Zuletzt aktualisiert: 15.07.2008.
    Theme
    Katalogfragen allgemein
  15. Hitchcock, S.; Bergmark, D.; Brody, T.; Gutteridge, C.; Carr, L.; Hall, W.; Lagoze, C.; Harnad, S.: Open citation linking : the way forward (2002) 0.01
    0.007878931 = product of:
      0.052526206 = sum of:
        0.017508736 = weight(_text_:software in 1207) [ClassicSimilarity], result of:
          0.017508736 = score(doc=1207,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.21915624 = fieldWeight in 1207, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1207)
        0.017508736 = weight(_text_:software in 1207) [ClassicSimilarity], result of:
          0.017508736 = score(doc=1207,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.21915624 = fieldWeight in 1207, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1207)
        0.017508736 = weight(_text_:software in 1207) [ClassicSimilarity], result of:
          0.017508736 = score(doc=1207,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.21915624 = fieldWeight in 1207, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1207)
      0.15 = coord(3/20)
    
    Abstract
    The speed of scientific communication - the rate of ideas affecting other researchers' ideas - is increasing dramatically. The factor driving this is free, unrestricted access to research papers. Measurements of user activity in mature eprint archives of research papers such as arXiv have shown, for the first time, the degree to which such services support an evolving network of texts commenting on, citing, classifying, abstracting, listing and revising other texts. The Open Citation project has built tools to measure this activity, to build new archives, and has been closely involved with the development of the infrastructure to support open access on which these new services depend. This is the story of the project, intertwined with the concurrent emergence of the Open Archives Initiative (OAI). The paper describes the broad scope of the project's work, showing how it has progressed from early demonstrators of reference linking to produce Citebase, a Web-based citation and impact-ranked search service, and how it has supported the development of the EPrints.org software for building OAI-compliant archives. The work has been underpinned by analysis and experiments on the semantics of documents (digital objects) to determine the features required for formally perfect linking - instantiated as an application programming interface (API) for reference linking - that will enable other applications to build on this work in broader digital library information environments.
  16. Hammond, T.; Hannay, T.; Lund, B.; Scott, J.: Social bookmarking tools (I) : a general review (2005) 0.01
    0.007799745 = product of:
      0.0519983 = sum of:
        0.017332766 = weight(_text_:software in 1188) [ClassicSimilarity], result of:
          0.017332766 = score(doc=1188,freq=4.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.21695362 = fieldWeight in 1188, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1188)
        0.017332766 = weight(_text_:software in 1188) [ClassicSimilarity], result of:
          0.017332766 = score(doc=1188,freq=4.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.21695362 = fieldWeight in 1188, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1188)
        0.017332766 = weight(_text_:software in 1188) [ClassicSimilarity], result of:
          0.017332766 = score(doc=1188,freq=4.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.21695362 = fieldWeight in 1188, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1188)
      0.15 = coord(3/20)
    
    Abstract
    Because, to paraphrase a pop music lyric from a certain rock and roll band of yesterday, "the Web is old, the Web is new, the Web is all, the Web is you", it seems like we might have to face up to some of these stark realities. With the introduction of new social software applications such as blogs, wikis, newsfeeds, social networks, and bookmarking tools (the subject of this paper), the claim that Shelley Powers makes in a Burningbird blog entry seems apposite: "This is the user's web now, which means it's my web and I can make the rules." Reinvention is revolution - it brings us always back to beginnings. We are here going to remind you of hyperlinks in all their glory, sell you on the idea of bookmarking hyperlinks, point you at other folks who are doing the same, and tell you why this is a good thing. Just as long as those hyperlinks (or let's call them plain old links) are managed, tagged, commented upon, and published onto the Web, they represent a user's own personal library placed on public record, which - when aggregated with other personal libraries - allows for rich, social networking opportunities. Why spill any ink (digital or not) in rewriting what someone else has already written about instead of just pointing at the original story and adding the merest of titles, descriptions and tags for future reference? More importantly, why not make these personal 'link playlists' available to oneself and to others from whatever browser or computer one happens to be using at the time? This paper reviews some current initiatives, as of early 2005, in providing public link management applications on the Web - utilities that are often referred to under the general moniker of 'social bookmarking tools'. There are a couple of things going on here: 1) server-side software aimed specifically at managing links with, crucially, a strong, social networking flavour, and 2) an unabashedly open and unstructured approach to tagging, or user classification, of those links.
  17. Krötzsch, M.; Hitzler, P.; Ehrig, M.; Sure, Y.: Category theory in ontology research : concrete gain from an abstract approach (2004 (?)) 0.01
    0.007716874 = product of:
      0.051445827 = sum of:
        0.017148608 = weight(_text_:23 in 4538) [ClassicSimilarity], result of:
          0.017148608 = score(doc=4538,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.23759183 = fieldWeight in 4538, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=4538)
        0.017148608 = weight(_text_:23 in 4538) [ClassicSimilarity], result of:
          0.017148608 = score(doc=4538,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.23759183 = fieldWeight in 4538, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=4538)
        0.017148608 = weight(_text_:23 in 4538) [ClassicSimilarity], result of:
          0.017148608 = score(doc=4538,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.23759183 = fieldWeight in 4538, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.046875 = fieldNorm(doc=4538)
      0.15 = coord(3/20)
    
    Date
    21.10.2018 15:23:19
  18. Hammond, T.; Hannay, T.; Lund, B.; Flack, M.: Social bookmarking tools (II) : a case study - Connotea (2005) 0.01
    0.006430729 = product of:
      0.042871524 = sum of:
        0.014290508 = weight(_text_:23 in 1189) [ClassicSimilarity], result of:
          0.014290508 = score(doc=1189,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.1979932 = fieldWeight in 1189, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1189)
        0.014290508 = weight(_text_:23 in 1189) [ClassicSimilarity], result of:
          0.014290508 = score(doc=1189,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.1979932 = fieldWeight in 1189, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1189)
        0.014290508 = weight(_text_:23 in 1189) [ClassicSimilarity], result of:
          0.014290508 = score(doc=1189,freq=2.0), product of:
            0.07217676 = queryWeight, product of:
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.02013827 = queryNorm
            0.1979932 = fieldWeight in 1189, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.5840597 = idf(docFreq=3336, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1189)
      0.15 = coord(3/20)
    
    Date
    26.12.2011 14:07:23
  19. Cohen, S.; Fereira, J.; Horne, A.; Kibbee, B.; Mistlebauer, H.; Smith, A.: MyLibrary : personalized electronic services in the Cornell University Library (2000) 0.01
    0.0063031456 = product of:
      0.04202097 = sum of:
        0.014006989 = weight(_text_:software in 1232) [ClassicSimilarity], result of:
          0.014006989 = score(doc=1232,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.17532499 = fieldWeight in 1232, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03125 = fieldNorm(doc=1232)
        0.014006989 = weight(_text_:software in 1232) [ClassicSimilarity], result of:
          0.014006989 = score(doc=1232,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.17532499 = fieldWeight in 1232, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03125 = fieldNorm(doc=1232)
        0.014006989 = weight(_text_:software in 1232) [ClassicSimilarity], result of:
          0.014006989 = score(doc=1232,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.17532499 = fieldWeight in 1232, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03125 = fieldNorm(doc=1232)
      0.15 = coord(3/20)
    
    Abstract
    Library users who are Web users expect customization and interactivity. MyLibrary is a Cornell University Library initiative to provide numerous personalized library services to Cornell University students, faculty, and staff. Currently, it consists of MyLinks, a tool for collecting and organizing resources for private use by a patron, and MyUpdates, a tool to help scholars stay informed of new resources provided by the library. This article provides an overview of the MyLibrary project, explains the rationale for the development of the service in the library, briefly discusses the hardware and software used for the service, and suggests some of the directions for future developments of the MyLibrary system. MyYahoo!, MyCNN, MyBookmarks, MyThis and MyThat. Internet users have demanded a personal face to the World Wide Web, and Web portals and information providers have responded. Why not MyLibrary? The Library and Information Technology Association (LITA) has defined MyLibrary-like services as the number one trend "worth keeping an eye on". "Library users who are Web users, a growing group," the experts agree, "expect customization, interactivity, and customer support. Approaches that are library-focused instead of user-focused will be increasingly irrelevant." In response to the needs of web-savvy patrons, the Cornell University Library (CUL) implemented a MyLibrary service this year, making finding and using library resources easier than ever. MyLibrary is an "umbrella" service for two new products: MyLinks and MyUpdates. Other products are in development. MyLibrary's MyLinks is a tool for collecting and organizing resources for private use by a patron. These resources may or may not be "official" Cornell University Library resources. Our patrons best understand this service as a "traveling set of bookmarks". Most patrons of the library use a variety of machines to access Internet resources. For example, you may have a computer at home and one at work. Why should you create your bookmarks twice, or carry around a diskette containing your bookmarks? Students who rely on lab computers never know which machine they will use next. With MyLinks, a patron's favorite sites are just a click away from any machine.
  20. Kubiszewski, I.; Cleveland, C.J.: ¬The Encyclopedia of Earth (2007) 0.01
    0.005515252 = product of:
      0.036768343 = sum of:
        0.012256115 = weight(_text_:software in 1170) [ClassicSimilarity], result of:
          0.012256115 = score(doc=1170,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.15340936 = fieldWeight in 1170, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1170)
        0.012256115 = weight(_text_:software in 1170) [ClassicSimilarity], result of:
          0.012256115 = score(doc=1170,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.15340936 = fieldWeight in 1170, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1170)
        0.012256115 = weight(_text_:software in 1170) [ClassicSimilarity], result of:
          0.012256115 = score(doc=1170,freq=2.0), product of:
            0.07989157 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02013827 = queryNorm
            0.15340936 = fieldWeight in 1170, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1170)
      0.15 = coord(3/20)
    
    Abstract
    The Encyclopedia of Earth (EoE) seeks to become the world's largest and most authoritative electronic source of information about the environments of Earth and their interactions with society. It is a free, fully searchable collection of articles written by scholars, professionals, educators, and experts who collaborate and review each other's work with oversight from an International Advisory Board. The articles are written in non-technical language and are available for free, with no commercial advertising to students, educators, scholars, professionals, decision makers, as well as to the general public. The scope of the Encyclopedia of Earth is the environment of the Earth broadly defined, with particular emphasis on the interaction between society and the natural spheres of the Earth. It will be built on the integrated knowledge from economists to philosophers to span all aspects of the environment. The Encyclopedia is being built bottom-up through the use of a wiki-software that allows users to freely create and edit content. New collaborations, ideas, and entries dynamically evolve in this environment. In this way, the Encyclopedia is a constantly evolving, self-organizing, expert-reviewed, and up-to-date source of environmental information. The motivation behind the Encyclopedia of Earth is simple. Go to GoogleT and type in climate change, pesticides, nuclear power, sustainable development, or any other important environmental issue. Doing so returns millions of results, some fraction of which are authoritative. The remainder is of poor or unknown quality.

Languages

  • e 64
  • d 14