Search (65 results, page 1 of 4)

  • × theme_ss:"Internet"
  • × type_ss:"el"
  1. Koch, T.; Ardö, A.: Automatic classification of full-text HTML-documents from one specific subject area : DESIRE II D3.6a, Working Paper 2 (2000) 0.09
    0.09283512 = product of:
      0.20887902 = sum of:
        0.06711562 = weight(_text_:applications in 1667) [ClassicSimilarity], result of:
          0.06711562 = score(doc=1667,freq=2.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.38913056 = fieldWeight in 1667, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.0625 = fieldNorm(doc=1667)
        0.014666359 = weight(_text_:of in 1667) [ClassicSimilarity], result of:
          0.014666359 = score(doc=1667,freq=6.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.23940048 = fieldWeight in 1667, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0625 = fieldNorm(doc=1667)
        0.03270373 = weight(_text_:systems in 1667) [ClassicSimilarity], result of:
          0.03270373 = score(doc=1667,freq=2.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.2716328 = fieldWeight in 1667, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.0625 = fieldNorm(doc=1667)
        0.09439332 = weight(_text_:software in 1667) [ClassicSimilarity], result of:
          0.09439332 = score(doc=1667,freq=6.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.6073436 = fieldWeight in 1667, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0625 = fieldNorm(doc=1667)
      0.44444445 = coord(4/9)
    
    Content
    1 Introduction / 2 Method overview / 3 Ei thesaurus preprocessing / 4 Automatic classification process: 4.1 Matching -- 4.2 Weighting -- 4.3 Preparation for display / 5 Results of the classification process / 6 Evaluations / 7 Software / 8 Other applications / 9 Experiments with universal classification systems / References / Appendix A: Ei classification service: Software / Appendix B: Use of the classification software as subject filter in a WWW harvester.
  2. Warnick, W.L.; Leberman, A.; Scott, R.L.; Spence, K.J.; Johnsom, L.A.; Allen, V.S.: Searching the deep Web : directed query engine applications at the Department of Energy (2001) 0.04
    0.0378924 = product of:
      0.1136772 = sum of:
        0.07118686 = weight(_text_:applications in 1215) [ClassicSimilarity], result of:
          0.07118686 = score(doc=1215,freq=4.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.41273528 = fieldWeight in 1215, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.046875 = fieldNorm(doc=1215)
        0.017962547 = weight(_text_:of in 1215) [ClassicSimilarity], result of:
          0.017962547 = score(doc=1215,freq=16.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.2932045 = fieldWeight in 1215, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=1215)
        0.0245278 = weight(_text_:systems in 1215) [ClassicSimilarity], result of:
          0.0245278 = score(doc=1215,freq=2.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.2037246 = fieldWeight in 1215, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.046875 = fieldNorm(doc=1215)
      0.33333334 = coord(3/9)
    
    Abstract
    Directed Query Engines, an emerging class of search engine specifically designed to access distributed resources on the deep web, offer the opportunity to create inexpensive digital libraries. Already, one such engine, Distributed Explorer, has been used to select and assemble high quality information resources and incorporate them into publicly available systems for the physical sciences. By nesting Directed Query Engines so that one query launches several other engines in a cascading fashion, enormous virtual collections may soon be assembled to form a comprehensive information infrastructure for the physical sciences. Once a Directed Query Engine has been configured for a set of information resources, distributed alerts tools can provide patrons with personalized, profile-based notices of recent additions to any of the selected resources. Due to the potentially enormous size and scope of Directed Query Engine applications, consideration must be given to issues surrounding the representation of large quantities of information from multiple, heterogeneous sources.
  3. Hitchcock, S.; Bergmark, D.; Brody, T.; Gutteridge, C.; Carr, L.; Hall, W.; Lagoze, C.; Harnad, S.: Open citation linking : the way forward (2002) 0.03
    0.031447157 = product of:
      0.094341464 = sum of:
        0.041947264 = weight(_text_:applications in 1207) [ClassicSimilarity], result of:
          0.041947264 = score(doc=1207,freq=2.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.2432066 = fieldWeight in 1207, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1207)
        0.018332949 = weight(_text_:of in 1207) [ClassicSimilarity], result of:
          0.018332949 = score(doc=1207,freq=24.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.2992506 = fieldWeight in 1207, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1207)
        0.034061253 = weight(_text_:software in 1207) [ClassicSimilarity], result of:
          0.034061253 = score(doc=1207,freq=2.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.21915624 = fieldWeight in 1207, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1207)
      0.33333334 = coord(3/9)
    
    Abstract
    The speed of scientific communication - the rate of ideas affecting other researchers' ideas - is increasing dramatically. The factor driving this is free, unrestricted access to research papers. Measurements of user activity in mature eprint archives of research papers such as arXiv have shown, for the first time, the degree to which such services support an evolving network of texts commenting on, citing, classifying, abstracting, listing and revising other texts. The Open Citation project has built tools to measure this activity, to build new archives, and has been closely involved with the development of the infrastructure to support open access on which these new services depend. This is the story of the project, intertwined with the concurrent emergence of the Open Archives Initiative (OAI). The paper describes the broad scope of the project's work, showing how it has progressed from early demonstrators of reference linking to produce Citebase, a Web-based citation and impact-ranked search service, and how it has supported the development of the EPrints.org software for building OAI-compliant archives. The work has been underpinned by analysis and experiments on the semantics of documents (digital objects) to determine the features required for formally perfect linking - instantiated as an application programming interface (API) for reference linking - that will enable other applications to build on this work in broader digital library information environments.
  4. Wilson, R.: ¬The role of ontologies in teaching and learning (2004) 0.03
    0.029994903 = product of:
      0.08998471 = sum of:
        0.04745791 = weight(_text_:applications in 3387) [ClassicSimilarity], result of:
          0.04745791 = score(doc=3387,freq=4.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.27515686 = fieldWeight in 3387, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03125 = fieldNorm(doc=3387)
        0.019401768 = weight(_text_:of in 3387) [ClassicSimilarity], result of:
          0.019401768 = score(doc=3387,freq=42.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.31669703 = fieldWeight in 3387, product of:
              6.4807405 = tf(freq=42.0), with freq of:
                42.0 = termFreq=42.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=3387)
        0.023125032 = weight(_text_:systems in 3387) [ClassicSimilarity], result of:
          0.023125032 = score(doc=3387,freq=4.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.19207339 = fieldWeight in 3387, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03125 = fieldNorm(doc=3387)
      0.33333334 = coord(3/9)
    
    Abstract
    Ontologies are currently a buzzword in many communities, hailed as a mechanism for making better use of the Web. They offer a shared definition of a domain that can be understood by computers, enabling them to complete more meaningful tasks. Although ontologies of different descriptions have been in development and use for some time, it is their potential as a key technology in the Semantic Web which is responsible for the current wave of interest. Communities have different expectations of the Semantic Web and how it will be realised, but it is generally believed that ontologies will play a major role. In light of their potential in this new context, much current effort is focusing an developing languages and tools. OWL (Web Ontology Language) has recently become a standard, and builds an top of existing Web languages such as XML and RDF to offer a high degree of expressiveness. A variety of tools are emerging for creating, editing and managing ontologies in OWL. Ontologies have a range of potential benefits and applications in further and higher education, including the sharing of information across educational systems, providing frameworks for learning object reuse, and enabling intelligent and personalised student support. The difficulties inherent in creating a model of a domain are being tackled, and the communities involved in ontology development are working together to achieve their vision of the Semantic Web. This Technology and Standards Watch report discusses ontologies and their role in the Semantic Web, with a special focus an their implications for teaching and learning. This report will introduce ontologies to the further and higher education community, explaining why they are being developed, what they hope to achieve, and their potential benefits to the community. Current ontology tools and standards will be described, and the emphasis will be an introducing the technology to a new audience and exploring its risks and potential applications in teaching and learning. At a time when educational programmes based an ontologies are starting to be developed, the author hopes to increase understanding of the key issues in the wider community.
    Content
    "Ontologies promise "a shared and common understanding of a domain that can be communicated between people and application systems" [1]. They attempt to formulate a thorough and rigorous representation of a domain by specifying all of its concepts, the relationships between them and the conditions and regulations of the domain. Ontologies can express hierarchical links between entities as well as other semantic relations. An example of part of an ontology is provided in Figure 1, in which it is specified not only that an author is a person and that a book is a publication, but also that an author writes a book and that a book has chapters."
  5. Klic, L.; Miller, M.; Nelson, J.K.; Germann, J.E.: Approaching the largest 'API' : extracting information from the Internet with Python (2018) 0.03
    0.025856502 = product of:
      0.11635426 = sum of:
        0.0089812735 = weight(_text_:of in 4239) [ClassicSimilarity], result of:
          0.0089812735 = score(doc=4239,freq=4.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.14660224 = fieldWeight in 4239, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=4239)
        0.107372984 = product of:
          0.21474597 = sum of:
            0.21474597 = weight(_text_:packages in 4239) [ClassicSimilarity], result of:
              0.21474597 = score(doc=4239,freq=6.0), product of:
                0.2706874 = queryWeight, product of:
                  6.9093957 = idf(docFreq=119, maxDocs=44218)
                  0.03917671 = queryNorm
                0.7933357 = fieldWeight in 4239, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  6.9093957 = idf(docFreq=119, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4239)
          0.5 = coord(1/2)
      0.22222222 = coord(2/9)
    
    Abstract
    This article explores the need for libraries to algorithmically access and manipulate the world's largest API: the Internet. The billions of pages on the 'Internet API' (HTTP, HTML, CSS, XPath, DOM, etc.) are easily accessible and manipulable. Libraries can assist in creating meaning through the datafication of information on the world wide web. Because most information is created for human consumption, some programming is required for automated extraction. Python is an easy-to-learn programming language with extensive packages and community support for web page automation. Four packages (Urllib, Selenium, BeautifulSoup, Scrapy) in Python can automate almost any web page for all sized projects. An example warrant data project is explained to illustrate how well Python packages can manipulate web pages to create meaning through assembling custom datasets.
  6. OWLED 2009; OWL: Experiences and Directions, Sixth International Workshop, Chantilly, Virginia, USA, 23-24 October 2009, Co-located with ISWC 2009. (2009) 0.02
    0.023978805 = product of:
      0.07193641 = sum of:
        0.043592874 = weight(_text_:applications in 3391) [ClassicSimilarity], result of:
          0.043592874 = score(doc=3391,freq=6.0), product of:
            0.17247584 = queryWeight, product of:
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.03917671 = queryNorm
            0.2527477 = fieldWeight in 3391, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.4025097 = idf(docFreq=1471, maxDocs=44218)
              0.0234375 = fieldNorm(doc=3391)
        0.010999769 = weight(_text_:of in 3391) [ClassicSimilarity], result of:
          0.010999769 = score(doc=3391,freq=24.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.17955035 = fieldWeight in 3391, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0234375 = fieldNorm(doc=3391)
        0.017343773 = weight(_text_:systems in 3391) [ClassicSimilarity], result of:
          0.017343773 = score(doc=3391,freq=4.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.14405504 = fieldWeight in 3391, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.0234375 = fieldNorm(doc=3391)
      0.33333334 = coord(3/9)
    
    Abstract
    The W3C OWL Web Ontology Language has been a W3C recommendation since 2004, and specification of its successor OWL 2 is being finalised. OWL plays an important role in an increasing number and range of applications and as experience using the language grows, new ideas for further extending its reach continue to be proposed. The OWL: Experiences and Direction (OWLED) workshop series is a forum for practitioners in industry and academia, tool developers, and others interested in OWL to describe real and potential applications, to share experience, and to discuss requirements for language extensions and modifications. The workshop will bring users, implementors and researchers together to measure the state of need against the state of the art, and to set an agenda for research and deployment in order to incorporate OWL-based technologies into new applications. This year's 2009 OWLED workshop will be co-located with the Eighth International Semantic Web Conference (ISWC), and the Third International Conference on Web Reasoning and Rule Systems (RR2009). It will be held in Chantilly, VA, USA on October 23 - 24, 2009. The workshop will concentrate on issues related to the development and W3C standardization of OWL 2, and beyond, but other issues related to OWL are also of interest, particularly those related to the task forces set up at OWLED 2007. As usual, the workshop will try to encourage participants to work together and will give space for discussions on various topics, to be decided and published at some point in the future. We ask participants to have a look at these topics and the accepted submissions before the workshop, and to prepare single "slides" that can be presented during these discussions. There will also be formal presentation of submissions to the workshop.
    Content
    Long Papers * Suggestions for OWL 3, Pascal Hitzler. * BestMap: Context-Aware SKOS Vocabulary Mappings in OWL 2, Rinke Hoekstra. * Mechanisms for Importing Modules, Bijan Parsia, Ulrike Sattler and Thomas Schneider. * A Syntax for Rules in OWL 2, Birte Glimm, Matthew Horridge, Bijan Parsia and Peter Patel-Schneider. * PelletSpatial: A Hybrid RCC-8 and RDF/OWL Reasoning and Query Engine, Markus Stocker and Evren Sirin. * The OWL API: A Java API for Working with OWL 2 Ontologies, Matthew Horridge and Sean Bechhofer. * From Justifications to Proofs for Entailments in OWL, Matthew Horridge, Bijan Parsia and Ulrike Sattler. * A Solution for the Man-Man Problem in the Family History Knowledge Base, Dmitry Tsarkov, Ulrike Sattler and Robert Stevens. * Towards Integrity Constraints in OWL, Evren Sirin and Jiao Tao. * Processing OWL2 ontologies using Thea: An application of logic programming, Vangelis Vassiliadis, Jan Wielemaker and Chris Mungall. * Reasoning in Metamodeling Enabled Ontologies, Nophadol Jekjantuk, Gerd Gröner and Jeff Z. Pan.
    Short Papers * A Database Backend for OWL, Jörg Henss, Joachim Kleb and Stephan Grimm. * Unifying SysML and OWL, Henson Graves. * The OWLlink Protocol, Thorsten Liebig, Marko Luther and Olaf Noppens. * A Reasoning Broker Framework for OWL, Juergen Bock, Tuvshintur Tserendorj, Yongchun Xu, Jens Wissmann and Stephan Grimm. * Change Representation For OWL 2 Ontologies, Raul Palma, Peter Haase, Oscar Corcho and Asunción Gómez-Pérez. * Practical Aspects of Query Rewriting for OWL 2, Héctor Pérez-Urbina, Ian Horrocks and Boris Motik. * CSage: Use of a Configurable Semantically Attributed Graph Editor as Framework for Editing and Visualization, Lawrence Levin. * A Conformance Test Suite for the OWL 2 RL/RDF Rules Language and the OWL 2 RDF-Based Semantics, Michael Schneider and Kai Mainzer. * Improving the Data Quality of Relational Databases using OBDA and OWL 2 QL, Olivier Cure. * Temporal Classes and OWL, Natalya Keberle. * Using Ontologies for Medical Image Retrieval - An Experiment, Jasmin Opitz, Bijan Parsia and Ulrike Sattler. * Task Representation and Retrieval in an Ontology-Guided Modelling System, Yuan Ren, Jens Lemcke, Andreas Friesen, Tirdad Rahmani, Srdjan Zivkovic, Boris Gregorcic, Andreas Bartho, Yuting Zhao and Jeff Z. Pan. * A platform for reasoning with OWL-EL knowledge bases in a Peer-to-Peer environment, Alexander De Leon and Michel Dumontier. * Axiomé: a Tool for the Elicitation and Management of SWRL Rules, Saeed Hassanpour, Martin O'Connor and Amar Das. * SQWRL: A Query Language for OWL, Martin O'Connor and Amar Das. * Classifying ELH Ontologies In SQL Databases, Vincent Delaitre and Yevgeny Kazakov. * A Semantic Web Approach to Represent and Retrieve Information in a Corporate Memory, Ana B. Rios-Alvarado, R. Carolina Medina-Ramirez and Ricardo Marcelin-Jimenez. * Towards a Graphical Notation for OWL 2, Elisa Kendall, Roy Bell, Roger Burkhart, Mark Dutra and Evan Wallace.
    Demo/Position Papers * Conjunctive Query Answering in Distributed Ontology Systems for Ontologies with Large OWL ABoxes, Xueying Chen and Michel Dumontier. * Node-Link and Containment Methods in Ontology Visualization, Julia Dmitrieva and Fons J. Verbeek. * A JC3IEDM OWL-DL Ontology, Steven Wartik. * Semantically Enabled Temporal Reasoning in a Virtual Observatory, Patrick West, Eric Rozell, Stephan Zednik, Peter Fox and Deborah L. McGuinness. * Developing an Ontology from the Application Up, James Malone, Tomasz Adamusiak, Ele Holloway, Misha Kapushesky and Helen Parkinson.
  7. Robbio, A. de; Maguolo, D.; Marini, A.: Scientific and general subject classifications in the digital world (2001) 0.02
    0.019814115 = product of:
      0.05944234 = sum of:
        0.015841477 = weight(_text_:of in 2) [ClassicSimilarity], result of:
          0.015841477 = score(doc=2,freq=28.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.25858206 = fieldWeight in 2, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=2)
        0.016351866 = weight(_text_:systems in 2) [ClassicSimilarity], result of:
          0.016351866 = score(doc=2,freq=2.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.1358164 = fieldWeight in 2, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03125 = fieldNorm(doc=2)
        0.027249003 = weight(_text_:software in 2) [ClassicSimilarity], result of:
          0.027249003 = score(doc=2,freq=2.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.17532499 = fieldWeight in 2, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03125 = fieldNorm(doc=2)
      0.33333334 = coord(3/9)
    
    Abstract
    In the present work we discuss opportunities, problems, tools and techniques encountered when interconnecting discipline-specific subject classifications, primarily organized as search devices in bibliographic databases, with general classifications originally devised for book shelving in public libraries. We first state the fundamental distinction between topical (or subject) classifications and object classifications. Then we trace the structural limitations that have constrained subject classifications since their library origins, and the devices that were used to overcome the gap with genuine knowledge representation. After recalling some general notions on structure, dynamics and interferences of subject classifications and of the objects they refer to, we sketch a synthetic overview on discipline-specific classifications in Mathematics, Computing and Physics, on one hand, and on general classifications on the other. In this setting we present The Scientific Classifications Page, which collects groups of Web pages produced by a pool of software tools for developing hypertextual presentations of single or paired subject classifications from sequential source files, as well as facilities for gathering information from KWIC lists of classification descriptions. Further we propose a concept-oriented methodology for interconnecting subject classifications, with the concrete support of a relational analysis of the whole Mathematics Subject Classification through its evolution since 1959. Finally, we recall a very basic method for interconnection provided by coreference in bibliographic records among index elements from different systems, and point out the advantages of establishing the conditions of a more widespread application of such a method. A part of these contents was presented under the title Mathematics Subject Classification and related Classifications in the Digital World at the Eighth International Conference Crimea 2001, "Libraries and Associations in the Transient World: New Technologies and New Forms of Cooperation", Sudak, Ukraine, June 9-17, 2001, in a special session on electronic libraries, electronic publishing and electronic information in science chaired by Bernd Wegner, Editor-in-Chief of Zentralblatt MATH.
  8. Networked knowledge organization systems (2001) 0.01
    0.014635116 = product of:
      0.06585802 = sum of:
        0.016802425 = weight(_text_:of in 6473) [ClassicSimilarity], result of:
          0.016802425 = score(doc=6473,freq=14.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.2742677 = fieldWeight in 6473, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=6473)
        0.0490556 = weight(_text_:systems in 6473) [ClassicSimilarity], result of:
          0.0490556 = score(doc=6473,freq=8.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.4074492 = fieldWeight in 6473, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.046875 = fieldNorm(doc=6473)
      0.22222222 = coord(2/9)
    
    Abstract
    Knowledge Organization Systems can comprise thesauri and other controlled lists of keywords, ontologies, classification systems, clustering approaches, taxonomies, gazetteers, dictionaries, lexical databases, concept maps/spaces, semantic road maps, etc. These schemas enable knowledge structuring and management, knowledge-based data processing and systematic access to knowledge structures in individual collections and digital libraries. Used as interactive information services on the Internet they have an increased potential to support the description, discovery and retrieval of heterogeneous information resources and to contribute to an overall resource discovery infrastructure
    Content
    This issue of the Journal of Digital Information evolved from a workshop on Networked Knowledge Organization Systems (NKOS) held at the Fourth European Conference on Research and Advanced Technology for Digital Libraries (ECDL2000) in Lisbon during September 2000. The focus of the workshop was European NKOS initiatives and projects and options for global cooperation. Workshop organizers were Martin Doerr, Traugott Koch, Dougles Tudhope and Repke de Vries. This group has, with Traugott Koch as the main editor and with the help of Linda Hill, cooperated in the editorial tasks for this special issue
    Source
    Journal of digital information. 1(2001) no.8
  9. Gore, E.; Bitta, M.D.; Cohen, D.: ¬The Digital Public Library of America and the National Digital Platform (2017) 0.01
    0.014363495 = product of:
      0.064635724 = sum of:
        0.023762217 = weight(_text_:of in 3655) [ClassicSimilarity], result of:
          0.023762217 = score(doc=3655,freq=28.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.38787308 = fieldWeight in 3655, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.046875 = fieldNorm(doc=3655)
        0.040873505 = weight(_text_:software in 3655) [ClassicSimilarity], result of:
          0.040873505 = score(doc=3655,freq=2.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.2629875 = fieldWeight in 3655, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=3655)
      0.22222222 = coord(2/9)
    
    Abstract
    The Digital Public Library of America brings together the riches of America's libraries, archives, and museums, and makes them freely available to the world. In order to do this, DPLA has had to build elements of the national digital platform to connect to those institutions and to serve their digitized materials to audiences. In this article, we detail the construction of two critical elements of our work: the decentralized national network of "hubs," which operate in states across the country; and a version of the Hydra repository software that is tailored to the needs of our community. This technology and the organizations that make use of it serve as the foundation of the future of DPLA and other projects that seek to take advantage of the national digital platform.
    Object
    Digital Public Library of America
  10. Choo, C.W.; Detlor, B.; Turnbull, D.: Information seeking on the Web : an integrated model of browsing and searching (2000) 0.01
    0.0138898 = product of:
      0.0625041 = sum of:
        0.014818345 = weight(_text_:of in 4438) [ClassicSimilarity], result of:
          0.014818345 = score(doc=4438,freq=8.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.24188137 = fieldWeight in 4438, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4438)
        0.047685754 = weight(_text_:software in 4438) [ClassicSimilarity], result of:
          0.047685754 = score(doc=4438,freq=2.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.30681872 = fieldWeight in 4438, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4438)
      0.22222222 = coord(2/9)
    
    Abstract
    This paper presents findings from a study of how knowledge workers use the Web to seek external information as part of their daily work. 34 users from 7 companies took part in the study. Participants were mainly IT-specialists, managers, and research/marketing/consulting staff working in organizations that included a large utility company; a major bank, and a consulting firm. Participants answered a detailed questionnaire and were interviewed individually in order to understand their information needs and information seeking preferences. A custom-developed WebTracker software application was installed on each of their work place PCs, and participants' Web-use activities were then recorded continuously during two-week periods
  11. Koch, T.; Ardö, A.; Brümmer, A.: ¬The building and maintenance of robot based internet search services : A review of current indexing and data collection methods. Prepared to meet the requirements of Work Package 3 of EU Telematics for Research, project DESIRE. Version D3.11v0.3 (Draft version 3) (1996) 0.01
    0.013267772 = product of:
      0.059704974 = sum of:
        0.021169065 = weight(_text_:of in 1669) [ClassicSimilarity], result of:
          0.021169065 = score(doc=1669,freq=50.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.34554482 = fieldWeight in 1669, product of:
              7.071068 = tf(freq=50.0), with freq of:
                50.0 = termFreq=50.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=1669)
        0.03853591 = weight(_text_:software in 1669) [ClassicSimilarity], result of:
          0.03853591 = score(doc=1669,freq=4.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.24794699 = fieldWeight in 1669, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03125 = fieldNorm(doc=1669)
      0.22222222 = coord(2/9)
    
    Abstract
    After a short outline of problems, possibilities and difficulties of systematic information retrieval on the Internet and a description of efforts for development in this area, a specification of the terminology for this report is required. Although the process of retrieval is generally seen as an iterative process of browsing and information retrieval and several important services on the net have taken this fact into consideration, the emphasis of this report lays on the general retrieval tools for the whole of Internet. In order to be able to evaluate the differences, possibilities and restrictions of the different services it is necessary to begin with organizing the existing varieties in a typological/ taxonomical survey. The possibilities and weaknesses will be briefly compared and described for the most important services in the categories robot-based WWW-catalogues of different types, list- or form-based catalogues and simultaneous or collected search services respectively. It will however for different reasons not be possible to rank them in order of "best" services. Still more important are the weaknesses and problems common for all attempts of indexing the Internet. The problems of the quality of the input, the technical performance and the general problem of indexing virtual hypertext are shown to be at least as difficult as the different aspects of harvesting, indexing and information retrieval. Some of the attempts made in the area of further development of retrieval services will be mentioned in relation to descriptions of the contents of documents and standardization efforts. Internet harvesting and indexing technology and retrieval software is thoroughly reviewed. Details about all services and software are listed in analytical forms in Annex 1-3.
  12. Kubiszewski, I.; Cleveland, C.J.: ¬The Encyclopedia of Earth (2007) 0.01
    0.009807495 = product of:
      0.04413373 = sum of:
        0.020290855 = weight(_text_:of in 1170) [ClassicSimilarity], result of:
          0.020290855 = score(doc=1170,freq=60.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.33120972 = fieldWeight in 1170, product of:
              7.745967 = tf(freq=60.0), with freq of:
                60.0 = termFreq=60.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1170)
        0.023842877 = weight(_text_:software in 1170) [ClassicSimilarity], result of:
          0.023842877 = score(doc=1170,freq=2.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.15340936 = fieldWeight in 1170, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1170)
      0.22222222 = coord(2/9)
    
    Abstract
    The Encyclopedia of Earth (EoE) seeks to become the world's largest and most authoritative electronic source of information about the environments of Earth and their interactions with society. It is a free, fully searchable collection of articles written by scholars, professionals, educators, and experts who collaborate and review each other's work with oversight from an International Advisory Board. The articles are written in non-technical language and are available for free, with no commercial advertising to students, educators, scholars, professionals, decision makers, as well as to the general public. The scope of the Encyclopedia of Earth is the environment of the Earth broadly defined, with particular emphasis on the interaction between society and the natural spheres of the Earth. It will be built on the integrated knowledge from economists to philosophers to span all aspects of the environment. The Encyclopedia is being built bottom-up through the use of a wiki-software that allows users to freely create and edit content. New collaborations, ideas, and entries dynamically evolve in this environment. In this way, the Encyclopedia is a constantly evolving, self-organizing, expert-reviewed, and up-to-date source of environmental information. The motivation behind the Encyclopedia of Earth is simple. Go to GoogleT and type in climate change, pesticides, nuclear power, sustainable development, or any other important environmental issue. Doing so returns millions of results, some fraction of which are authoritative. The remainder is of poor or unknown quality.
    This illustrates a stark reality of the Web. There are many resources for environmental content, but there is no central repository of authoritative information that meets the needs of diverse user communities. The Encyclopedia of Earth aims to fill that niche by providing content that is both free and reliable. Still in its infancy, the EoE already is an integral part of the emerging effort to increase free and open access to trusted information on the Web. It is a trusted content source for authoritative indexes such as the Online Access to Research in the Environment Initiative, the Health InterNetwork Access to Research Initiative, the Open Education Resources Commons, Scirus, DLESE, WiserEarth, among others. Our initial Content Partners include the American Institute of Physics, the University of California Museum of Paleontology, TeacherServe®, the U.S. Geological Survey, the International Arctic Science Committee, the World Wildlife Fund, Conservation International, the Biodiversity Institute of Ontario, and the United Nations Environment Programme, to name just a few. The full partner list here can be found at <http://www.eoearth.org/article/Content_Partners>. We have a diversity of article types including standard subject articles, biographies, place-based entries, country profiles, and environmental classics. We recently launched our E-Book series, full-text, fully searchable books with internal hyperlinks to EoE articles. The eBooks include new releases by distinguished scholars as well as classics such as Walden and On the Origin of Species. Because history can be an important guide to the future, we have added an Environmental Classics section that includes such historical works as Energy from Fossil Fuels by M. King Hubbert and Undersea by Rachel Carson. Our services and features will soon be expanded. The EoE will soon be available in different languages giving a wider range of users access, users will be able to search it geographically or by a well-defined, expert created taxonomy, and teachers will be able to use the EoE to create unique curriculum for their courses.
    Content
    The home page for the Encyclopedia of Earth is located at <http://www.eoearth.org/>.
    Object
    Encyclopedia of Earth
  13. Zhang, A.: Multimedia file formats on the Internet : a beginner's guide for PC users (1995) 0.01
    0.009083001 = product of:
      0.08174701 = sum of:
        0.08174701 = weight(_text_:software in 3212) [ClassicSimilarity], result of:
          0.08174701 = score(doc=3212,freq=2.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.525975 = fieldWeight in 3212, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.09375 = fieldNorm(doc=3212)
      0.11111111 = coord(1/9)
    
    Abstract
    Darstellung der verschiedenen Dateiformate, wie sie im Internet verwendet werden sowie die Möglichkeiten, die Dateien zu nutzen (einschl. Angaben zu Software etc.)
  14. Van de Sompel, H.; Hochstenbach, P.: Reference linking in a hybrid library environment : part 1: frameworks for linking (1999) 0.01
    0.008659224 = product of:
      0.038966507 = sum of:
        0.015841477 = weight(_text_:of in 1244) [ClassicSimilarity], result of:
          0.015841477 = score(doc=1244,freq=28.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.25858206 = fieldWeight in 1244, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=1244)
        0.023125032 = weight(_text_:systems in 1244) [ClassicSimilarity], result of:
          0.023125032 = score(doc=1244,freq=4.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.19207339 = fieldWeight in 1244, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03125 = fieldNorm(doc=1244)
      0.22222222 = coord(2/9)
    
    Abstract
    The creation of services linking related information entities is an area that is attracting an ever increasing interest in the ongoing development of the World Wide Web in general, and of research-related information systems in particular. Currently, both practice and theory point at linking services as being a major domain for innovation enabled by digital communication of content. Publishers, subscription agents, researchers and libraries are all looking into ways to create added value by linking related information entities, as such presenting the information within a broader context estimated to be relevant to the users of the information. This is the first of two articles in D-Lib Magazine on this topic. This first part describes the current state-of-the-art and contrasts various approaches to the problem. It identifies static and dynamic linking solutions as well as open and closed linking frameworks. It also includes an extensive bibliography. The second part, SFX, a Generic Linking Solution describes a system that we have developed for linking in a hybrid working environment. The creation of services linking related information entities is an area that is attracting an ever increasing interest in the ongoing development of the World Wide Web in general, and of research-related information systems in particular. Although most writings on electronic scientific communication have touted other benefits, such as the increase in communication speed, the possibility to exchange multimedia content and the absence of limitations on the length of research papers, currently both practice and theory point at linking services as being a major opportunity for improved communication of content. Publishers, subscription agents, researchers and libraries are all looking into ways to create added-value by linking related information entities, as such presenting the information within a broader context estimated to be relevant to the users of the information.
  15. cis: Nationalbibliothek will das deutsche Internet kopieren (2008) 0.01
    0.007362602 = product of:
      0.033131707 = sum of:
        0.023842877 = weight(_text_:software in 4609) [ClassicSimilarity], result of:
          0.023842877 = score(doc=4609,freq=2.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.15340936 = fieldWeight in 4609, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4609)
        0.009288831 = product of:
          0.018577661 = sum of:
            0.018577661 = weight(_text_:22 in 4609) [ClassicSimilarity], result of:
              0.018577661 = score(doc=4609,freq=2.0), product of:
                0.13719016 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03917671 = queryNorm
                0.1354154 = fieldWeight in 4609, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=4609)
          0.5 = coord(1/2)
      0.22222222 = coord(2/9)
    
    Content
    "Ab heute müsste kopiert und abgespeichert werden - eigentlich. Fast jede Webseite, samt Bildern, Texten und sogar Audiodateien soll die Deutsche Nationalbibliothek archivieren, im Auftrag der Bundesregierung. Das Kopieren und Abspeichern sollen die Bundes-Bibliothekare aber nicht übernehmen. Das sollen die Seitenbetreiber machen. Am Donnerstag wurde die Verordnung über die Pflichtablieferung von Medienwerken an die Deutsche Nationalbibliothek im Bundesgesetzblatt veröffentlicht - damit ist sie rechtskräftig. Auf den Seiten der Deutschen Nationalbibliothek ist auch nachzulesen, was da von nun an alles gesammelt werden soll: "Alle Darstellungen in Schrift, Bild und Ton, die in öffentlichen Netzen zugänglich gemacht werden." Mit anderen Worten: Die Nationalbibliothek möchte in regelmäßigen Abständen vollständige Kopien des deutschen Internets bekommen, einschließlich multimedialer Elemente. Ausgenommen sind lediglich "zeitlich begrenzte Vorabveröffentlichungen, reine Software- oder Anwendungstools und auch Fernseh- und Hörfunkproduktionen" Der Branchenverband Bitkom hatte schon im vergangenen Jahr gegen das Gesetz protestiert, der Widerstand verpuffte jedoch ohne Konsequenz. Deutsche Firmen werde die tatsächliche Umsetzung der Regelung 115 Millionen Euro jährlich kosten, schätzte der Verband damals. Was an Kosten auf den Steuerzahler zukommt, ist noch vollkommen unklar. Die Regelung ist das Ergebnis eines Gesetzes, das schon aus dem Jahr 2006 stammt: Dem Gesetz über die Deutsche Nationalbibliothek ( DNBG). Jetzt ist das Gesetz samt erklärender Verordnung da - aber wie das alles genau funktionieren soll, wer die riesigen Mengen Speicherplatz zur Verfügung stehen wird, das scheint noch niemand so recht zu wissen.
    Date
    24.10.2008 14:19:22
  16. Noerr, P.: ¬The Digital Library Tool Kit (2001) 0.01
    0.007154076 = product of:
      0.03219334 = sum of:
        0.015841477 = weight(_text_:of in 6774) [ClassicSimilarity], result of:
          0.015841477 = score(doc=6774,freq=28.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.25858206 = fieldWeight in 6774, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=6774)
        0.016351866 = weight(_text_:systems in 6774) [ClassicSimilarity], result of:
          0.016351866 = score(doc=6774,freq=2.0), product of:
            0.12039685 = queryWeight, product of:
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03917671 = queryNorm
            0.1358164 = fieldWeight in 6774, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.0731742 = idf(docFreq=5561, maxDocs=44218)
              0.03125 = fieldNorm(doc=6774)
      0.22222222 = coord(2/9)
    
    Abstract
    This second edition is an update and expansion of the original April 1998 edition. It contains more of everything. In particular, the resources section has been expanded and updated. This document is designed to help those who are contemplating setting up a digital library. Whether this is a first time computerization effort or an extension of an existing library's services, there are questions to be answered, deci-sions to be made, and work to be done. This document covers all those stages and more. The first section (Chapter 1) is a series of questions to ask yourself and your organization. The questions are designed generally to raise issues rather than to provide definitive answers. The second section (Chapters 2-5) discusses the planning and implementation of a digital library. It raises some issues which are specific, and contains information to help answer the specifics and a host of other aspects of a digital li-brary project. The third section (Chapters 6 -7) includes resources and a look at current research, existing digital library systems, and the future. These chapters enable you to find additional resources and help, as well as show you where to look for interesting examples of the current state of the art
    Footnote
    This Digital Library Tool Kit was sponsored by Sun Microsystems, Inc. to address some of the leading questions that academic institutions, public libraries, government agencies, and museums face in trying to develop, manage, and distribute digital content. The evolution of Java programming, digital object standards, Internet access, electronic commerce, and digital media management models is causing educators, CIOs, and librarians to rethink many of their traditional goals and modes of operation. New audiences, continuous access to collections, and enhanced services to user communities are enabled. As one of the leading technology providers to education and library communities, Sun is pleased to present this comprehensive introduction to digital libraries
  17. Ding, J.: Can data die? : why one of the Internet's oldest images lives on wirhout its subjects's consent (2021) 0.01
    0.0067829583 = product of:
      0.030523311 = sum of:
        0.013492684 = weight(_text_:of in 423) [ClassicSimilarity], result of:
          0.013492684 = score(doc=423,freq=52.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.22024246 = fieldWeight in 423, product of:
              7.2111025 = tf(freq=52.0), with freq of:
                52.0 = termFreq=52.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.01953125 = fieldNorm(doc=423)
        0.017030627 = weight(_text_:software in 423) [ClassicSimilarity], result of:
          0.017030627 = score(doc=423,freq=2.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.10957812 = fieldWeight in 423, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.01953125 = fieldNorm(doc=423)
      0.22222222 = coord(2/9)
    
    Abstract
    In 2021, sharing content is easier than ever. Our lingua franca is visual: memes, infographics, TikToks. Our references cross borders and platforms, shared and remixed a hundred different ways in minutes. Digital culture is collective by default and has us together all around the world. But as the internet reaches its "dirty 30s," what happens when pieces of digital culture that have been saved, screenshotted, and reposted for years need to retire? Let's dig into the story of one of these artifacts: The Lenna image. The Lenna image may be relatively unknown in pop culture today, but in the engineering world, it remains an icon. I first encountered the image in an undergrad class, then grad school, and then all over the sites and software I use every day as a tech worker like Github, OpenCV, Stack Overflow, and Quora. To understand where the image is today, you have to understand how it got here. So, I decided to scrape Google scholar, search, and reverse image search results to track down thousands of instances of the image across the internet (see more in the methods section).
    Lena Forsén, the real human behind the Lenna image, was first published in Playboy in 1972. Soon after, USC engineers searching for a suitable test image for their image processing research sought inspiration from the magazine. They deemed Lenna the right fit and scanned the image into digital, RGB existence. From here, the story of the image follows the story of the internet. Lenna was one of the first inhabitants of ARPANet, the internet's predecessor, and then the world wide web. While the image's reach was limited to a few research papers in the '70s and '80s, in 1991, Lenna was featured on the cover of an engineering journal alongside another popular test image, Peppers. This caught the attention of Playboy, which threatened a copyright infringement lawsuit. Engineers who had grown attached to Lenna fought back. Ultimately, they prevailed, and as a Playboy VP reflected on the drama: "We decided we should exploit this because it is a phenomenon." The Playboy controversy canonized Lenna in engineering folklore and prompted an explosion of conversation about the image. Image hits on the internet rose to a peak number in 1995.
    In the 21st century, the image has remained a common sight in classrooms and on TV, including a feature on Silicon Valley in 2014. Pushback towards the use of the image also grew in the 2010s leading up to 2019, when the Losing Lena documentary was released. Forsén shares her side of the story and asks for her image to be retired: "I retired from modelling a long time ago. It's time I retired from tech, too. We can make a simple change today that creates a lasting change for tomorrow. Let's commit to losing me." After the film's release, many of my female colleagues shared stories about their own encounters with the image throughout their careers. When one of the only women this well referenced, respected, and remembered in your field is known for a nude photo that was taken of her and is now used without her consent, it inevitably shapes the perception of the position of women in tech and the value of our contributions. The film called on the engineering community to stop their spread of the image and use alternatives instead. This led to efforts to remove the image from textbooks and production code and a slow, but noticeable decline in the image's use for research.
    But despite this progress, almost 2 years later, the use of Lenna continues. The image appears on the internet in 30+ different languages in the last decade, including 10+ languages in 2021. The image's spread across digital geographies has mirrored this geographical growth, moving from mostly .org domains before 1990 to over 100 different domains today, notably .com and .edu, along with others. Within the .edu world, the Lenna image continues to appear in homework questions, class slides and to be hosted on educational and research sites, ensuring that it is passed down to new generations of engineers. Whether it's due to institutional negligence or defiance, it seems that for now, the image is here to stay.
    Content
    "Having known Lenna for almost a decade, I have struggled to understand what the story of the image means for what tech culture is and what it is becoming. To me, the crux of the Lenna story is how little power we have over our data and how it is used and abused. This threat seems disproportionately higher for women who are often overrepresented in internet content, but underrepresented in internet company leadership and decision making. Given this reality, engineering and product decisions will continue to consciously (and unconsciously) exclude our needs and concerns. While social norms are changing towards non-consensual data collection and data exploitation, digital norms seem to be moving in the opposite direction. Advancements in machine learning algorithms and data storage capabilities are only making data misuse easier. Whether the outcome is revenge porn or targeted ads, surveillance or discriminatory AI, if we want a world where our data can retire when it's outlived its time, or when it's directly harming our lives, we must create the tools and policies that empower data subjects to have a say in what happens to their data. including allowing their data to die."
  18. Blosser, J.; Michaelson, R.; Routh. R.; Xia, P.: Defining the landscape of Web resources : Concluding Report of the BAER Web Resources Sub-Group (2000) 0.01
    0.0064601265 = product of:
      0.02907057 = sum of:
        0.018454762 = weight(_text_:of in 1447) [ClassicSimilarity], result of:
          0.018454762 = score(doc=1447,freq=38.0), product of:
            0.061262865 = queryWeight, product of:
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03917671 = queryNorm
            0.30123898 = fieldWeight in 1447, product of:
              6.164414 = tf(freq=38.0), with freq of:
                38.0 = termFreq=38.0
              1.5637573 = idf(docFreq=25162, maxDocs=44218)
              0.03125 = fieldNorm(doc=1447)
        0.010615807 = product of:
          0.021231614 = sum of:
            0.021231614 = weight(_text_:22 in 1447) [ClassicSimilarity], result of:
              0.021231614 = score(doc=1447,freq=2.0), product of:
                0.13719016 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03917671 = queryNorm
                0.15476047 = fieldWeight in 1447, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1447)
          0.5 = coord(1/2)
      0.22222222 = coord(2/9)
    
    Abstract
    The BAER Web Resources Group was charged in October 1999 with defining and describing the parameters of electronic resources that do not clearly belong to the categories being defined by the BAER Digital Group or the BAER Electronic Journals Group. After some difficulty identifying precisely which resources fell under the Group's charge, we finally named the following types of resources for our consideration: web sites, electronic texts, indexes, databases and abstracts, online reference resources, and networked and non-networked CD-ROMs. Electronic resources are a vast and growing collection that touch nearly every department within the Library. It is unrealistic to think one department can effectively administer all aspects of the collection. The Group then began to focus on the concern of bibliographic access to these varied resources, and to define parameters for handling or processing them within the Library. Some key elements became evident as the work progressed. * Selection process of resources to be acquired for the collection * Duplication of effort * Use of CORC * Resource Finder design * Maintenance of Resource Finder * CD-ROMs not networked * Communications * Voyager search limitations. An unexpected collaboration with the Web Development Committee on the Resource Finder helped to steer the Group to more detailed descriptions of bibliographic access. This collaboration included development of data elements for the Resource Finder database, and some discussions on Library staff processing of the resources. The Web Resources Group invited expert testimony to help the Group broaden its view to envision public use of the resources and discuss concerns related to technical services processing. The first testimony came from members of the Resource Finder Committee. Some background information on the Web Development Resource Finder Committee was shared. The second testimony was from librarians who select electronic texts. Three main themes were addressed: accessing CD-ROMs; the issue of including non-networked CD-ROMs in the Resource Finder; and, some special concerns about electronic texts. The third testimony came from librarians who select indexes and abstracts and also provide Reference services. Appendices to this report include minutes of the meetings with the experts (Appendix A), a list of proposed data elements to be used in the Resource Finder (Appendix B), and recommendations made to the Resource Finder Committee (Appendix C). Below are summaries of the key elements.
    Date
    21. 4.2002 10:22:31
  19. Ask me[@sk.me]: your global information guide : der Wegweiser durch die Informationswelten (1996) 0.00
    0.0047181365 = product of:
      0.042463228 = sum of:
        0.042463228 = product of:
          0.084926456 = sum of:
            0.084926456 = weight(_text_:22 in 5837) [ClassicSimilarity], result of:
              0.084926456 = score(doc=5837,freq=2.0), product of:
                0.13719016 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03917671 = queryNorm
                0.61904186 = fieldWeight in 5837, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.125 = fieldNorm(doc=5837)
          0.5 = coord(1/2)
      0.11111111 = coord(1/9)
    
    Date
    30.11.1996 13:22:37
  20. Wetzel, D.: Bücher, Spielzeug, Daten, Krieg : Die globalen Software- und Internetriesen drängen in das Geschäft mit »Sicherheit«. (2019) 0.00
    0.0045415005 = product of:
      0.040873505 = sum of:
        0.040873505 = weight(_text_:software in 3955) [ClassicSimilarity], result of:
          0.040873505 = score(doc=3955,freq=2.0), product of:
            0.15541996 = queryWeight, product of:
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.03917671 = queryNorm
            0.2629875 = fieldWeight in 3955, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.9671519 = idf(docFreq=2274, maxDocs=44218)
              0.046875 = fieldNorm(doc=3955)
      0.11111111 = coord(1/9)
    

Years

Languages

  • e 52
  • d 11
  • el 1
  • More… Less…

Types