Search (449 results, page 1 of 23)

  • × type_ss:"el"
  1. XFML Core - eXchangeable Faceted Metadata Language (2003) 0.12
    0.11721319 = product of:
      0.23442638 = sum of:
        0.20071624 = weight(_text_:markup in 6673) [ClassicSimilarity], result of:
          0.20071624 = score(doc=6673,freq=2.0), product of:
            0.27638784 = queryWeight, product of:
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.042049456 = queryNorm
            0.72621226 = fieldWeight in 6673, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.078125 = fieldNorm(doc=6673)
        0.033710144 = product of:
          0.10113043 = sum of:
            0.10113043 = weight(_text_:language in 6673) [ClassicSimilarity], result of:
              0.10113043 = score(doc=6673,freq=4.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.6130145 = fieldWeight in 6673, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.078125 = fieldNorm(doc=6673)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    The specification for XFML, a markup language designed to handle faceted classifications. Browsing the site (http://www.xfml.org/) will reveal news about XFML and links to related software and web sites. XFML is not an officially recognized Internet standard, but is the de facto standard.
  2. Shah, U.; Finin, T.; Joshi, A.; Cost, R.S.; Mayfield, J.: Information retrieval on the Semantic Web (2002) 0.11
    0.10769231 = product of:
      0.21538462 = sum of:
        0.19869895 = weight(_text_:markup in 696) [ClassicSimilarity], result of:
          0.19869895 = score(doc=696,freq=4.0), product of:
            0.27638784 = queryWeight, product of:
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.042049456 = queryNorm
            0.7189135 = fieldWeight in 696, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.0546875 = fieldNorm(doc=696)
        0.01668567 = product of:
          0.05005701 = sum of:
            0.05005701 = weight(_text_:language in 696) [ClassicSimilarity], result of:
              0.05005701 = score(doc=696,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.30342668 = fieldWeight in 696, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=696)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    We describe an apporach to retrieval of documents that consist of both free text and semantically enriched markup. In particular, we present the design and implementation prototype of a framework in which both documents and queries can be marked up with statements in the DAML+OIL semantic web language. These statement provide both structured and semi-structured information about the documents and their content. We claim that indexing text and semantic markup will significantly improve retrieval performance. Outr approach allows inferencing to be done over this information at several points: when a document is indexed,when a query is processed and when query results are evaluated.
  3. Powell, J.; Fox, E.A.: Multilingual federated searching across heterogeneous collections (1998) 0.09
    0.08982116 = product of:
      0.17964232 = sum of:
        0.16057299 = weight(_text_:markup in 1250) [ClassicSimilarity], result of:
          0.16057299 = score(doc=1250,freq=2.0), product of:
            0.27638784 = queryWeight, product of:
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.042049456 = queryNorm
            0.5809698 = fieldWeight in 1250, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.0625 = fieldNorm(doc=1250)
        0.019069336 = product of:
          0.05720801 = sum of:
            0.05720801 = weight(_text_:language in 1250) [ClassicSimilarity], result of:
              0.05720801 = score(doc=1250,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.34677336 = fieldWeight in 1250, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1250)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    This article describes a scalable system for searching heterogeneous multilingual collections on the World Wide Web. It details a markup language for describing the characteristics of a search engine and its interface, and a protocol for requesting word translations between languages.
  4. Bechhofer, S.; Harmelen, F. van; Hendler, J.; Horrocks, I.; McGuinness, D.L.; Patel-Schneider, P.F.; Stein, L.A.: OWL Web Ontology Language Reference (2004) 0.09
    0.088905826 = product of:
      0.17781165 = sum of:
        0.14050137 = weight(_text_:markup in 4684) [ClassicSimilarity], result of:
          0.14050137 = score(doc=4684,freq=2.0), product of:
            0.27638784 = queryWeight, product of:
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.042049456 = queryNorm
            0.5083486 = fieldWeight in 4684, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4684)
        0.037310295 = product of:
          0.111930884 = sum of:
            0.111930884 = weight(_text_:language in 4684) [ClassicSimilarity], result of:
              0.111930884 = score(doc=4684,freq=10.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.6784828 = fieldWeight in 4684, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4684)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    The Web Ontology Language OWL is a semantic markup language for publishing and sharing ontologies on the World Wide Web. OWL is developed as a vocabulary extension of RDF (the Resource Description Framework) and is derived from the DAML+OIL Web Ontology Language. This document contains a structured informal description of the full set of OWL language constructs and is meant to serve as a reference for OWL users who want to construct OWL ontologies.
  5. Bartolo, L.M.; Lowe, C.S.; Sadoway, D.R.; Powell, A.C.; Glotzer, S.C.: NSDL MatDL : exploring digital library roles (2005) 0.08
    0.07939145 = product of:
      0.1587829 = sum of:
        0.14192782 = weight(_text_:markup in 1181) [ClassicSimilarity], result of:
          0.14192782 = score(doc=1181,freq=4.0), product of:
            0.27638784 = queryWeight, product of:
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.042049456 = queryNorm
            0.51350963 = fieldWeight in 1181, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1181)
        0.016855072 = product of:
          0.050565217 = sum of:
            0.050565217 = weight(_text_:language in 1181) [ClassicSimilarity], result of:
              0.050565217 = score(doc=1181,freq=4.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.30650726 = fieldWeight in 1181, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1181)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    A primary goal of the NSDL Materials Digital Library (MatDL) is to bring materials science research and education closer together. MatDL is exploring the various roles digital libraries can serve in the materials science community including: 1) supporting a virtual lab, 2) developing markup language applications, and 3) building tools for metadata capture. MatDL is being integrated into an MIT virtual laboratory experience. Early student self-assessment survey results expressed positive opinions of the potential value of MatDL in supporting a virtual lab and in accomplishing additional educational objectives. A separate survey suggested that the effectiveness of a virtual lab may approach that of a physical lab on some laboratory learning objectives. MatDL is collaboratively developing a materials property grapher (KSU and MIT) and a submission tool (KSU and U-M). MatML is an extensible markup language for exchanging materials information developed by materials data experts in industry, government, standards organizations, and professional societies. The web-based MatML grapher allows students to compare selected materials properties across approximately 80 MatML-tagged materials. The MatML grapher adds value in this educational context by allowing students to utilize real property data to make optimal material selection decisions. The submission tool has been integrated into the regular workflow of U-M students and researchers generating nanostructure images. It prompts users for domain-specific information, automatically generating and attaching keywords and editable descriptions.
  6. Mayfield, J.; Finin, T.: Information retrieval on the Semantic Web : integrating inference and retrieval 0.08
    0.07689732 = product of:
      0.15379465 = sum of:
        0.14050137 = weight(_text_:markup in 4330) [ClassicSimilarity], result of:
          0.14050137 = score(doc=4330,freq=2.0), product of:
            0.27638784 = queryWeight, product of:
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.042049456 = queryNorm
            0.5083486 = fieldWeight in 4330, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4330)
        0.013293281 = product of:
          0.039879844 = sum of:
            0.039879844 = weight(_text_:22 in 4330) [ClassicSimilarity], result of:
              0.039879844 = score(doc=4330,freq=2.0), product of:
                0.14725003 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.042049456 = queryNorm
                0.2708308 = fieldWeight in 4330, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4330)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    One vision of the Semantic Web is that it will be much like the Web we know today, except that documents will be enriched by annotations in machine understandable markup. These annotations will provide metadata about the documents as well as machine interpretable statements capturing some of the meaning of document content. We discuss how the information retrieval paradigm might be recast in such an environment. We suggest that retrieval can be tightly bound to inference. Doing so makes today's Web search engines useful to Semantic Web inference engines, and causes improvements in either retrieval or inference to lead directly to improvements in the other.
    Date
    12. 2.2011 17:35:22
  7. What is Schema.org? (2011) 0.07
    0.07374786 = product of:
      0.29499143 = sum of:
        0.29499143 = weight(_text_:markup in 4437) [ClassicSimilarity], result of:
          0.29499143 = score(doc=4437,freq=12.0), product of:
            0.27638784 = queryWeight, product of:
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.042049456 = queryNorm
            1.0673097 = fieldWeight in 4437, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.046875 = fieldNorm(doc=4437)
      0.25 = coord(1/4)
    
    Abstract
    This site provides a collection of schemas, i.e., html tags, that webmasters can use to markup their pages in ways recognized by major search providers. Search engines including Bing, Google and Yahoo! rely on this markup to improve the display of search results, making it easier for people to find the right web pages. Many sites are generated from structured data, which is often stored in databases. When this data is formatted into HTML, it becomes very difficult to recover the original structured data. Many applications, especially search engines, can benefit greatly from direct access to this structured data. On-page markup enables search engines to understand the information on web pages and provide richer search results in order to make it easier for users to find relevant information on the web. Markup can also enable new tools and applications that make use of the structure. A shared markup vocabulary makes easier for webmasters to decide on a markup schema and get the maximum benefit for their efforts. So, in the spirit of sitemaps.org, Bing, Google and Yahoo! have come together to provide a shared collection of schemas that webmasters can use.
  8. Lee, M.; Baillie, S.; Dell'Oro, J.: TML: a Thesaural Markpup Language (200?) 0.07
    0.072600774 = product of:
      0.14520155 = sum of:
        0.12042975 = weight(_text_:markup in 1622) [ClassicSimilarity], result of:
          0.12042975 = score(doc=1622,freq=2.0), product of:
            0.27638784 = queryWeight, product of:
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.042049456 = queryNorm
            0.43572736 = fieldWeight in 1622, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.046875 = fieldNorm(doc=1622)
        0.0247718 = product of:
          0.0743154 = sum of:
            0.0743154 = weight(_text_:language in 1622) [ClassicSimilarity], result of:
              0.0743154 = score(doc=1622,freq=6.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.45047188 = fieldWeight in 1622, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1622)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    Thesauri are used to provide controlled vocabularies for resource classification. Their use can greatly assist document discovery because thesauri man date a consistent shared terminology for describing documents. A particular thesauras classifies documents according to an information community's needs. As a result, there are many different thesaural schemas. This has led to a proliferation of schema-specific thesaural systems. In our research, we exploit schematic regularities to design a generic thesaural ontology and specfiy it as a markup language. The language provides a common representational framework in which to encode the idiosyncrasies of specific thesauri. This approach has several advantages: it offers consistent syntax and semantics in which to express thesauri; it allows general purpose thesaural applications to leverage many thesauri; and it supports a single thesaural user interface by which information communities can consistently organise, score and retrieve electronic documents.
  9. Mayo, D.; Bowers, K.: ¬The devil's shoehorn : a case study of EAD to ArchivesSpace migration at a large university (2017) 0.07
    0.07167879 = product of:
      0.14335757 = sum of:
        0.10035812 = weight(_text_:markup in 3373) [ClassicSimilarity], result of:
          0.10035812 = score(doc=3373,freq=2.0), product of:
            0.27638784 = queryWeight, product of:
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.042049456 = queryNorm
            0.36310613 = fieldWeight in 3373, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3373)
        0.042999458 = product of:
          0.064499184 = sum of:
            0.03575501 = weight(_text_:language in 3373) [ClassicSimilarity], result of:
              0.03575501 = score(doc=3373,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.21673335 = fieldWeight in 3373, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3373)
            0.028744178 = weight(_text_:29 in 3373) [ClassicSimilarity], result of:
              0.028744178 = score(doc=3373,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.19432661 = fieldWeight in 3373, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3373)
          0.6666667 = coord(2/3)
      0.5 = coord(2/4)
    
    Abstract
    A band of archivists and IT professionals at Harvard took on a project to convert nearly two million descriptions of archival collection components from marked-up text into the ArchivesSpace archival metadata management system. Starting in the mid-1990s, Harvard was an alpha implementer of EAD, an SGML (later XML) text markup language for electronic inventories, indexes, and finding aids that archivists use to wend their way through the sometimes quirky filing systems that bureaucracies establish for their records or the utter chaos in which some individuals keep their personal archives. These pathfinder documents, designed to cope with messy reality, can themselves be difficult to classify. Portions of them are rigorously structured, while other parts are narrative. Early documents predate the establishment of the standard; many feature idiosyncratic encoding that had been through several machine conversions, while others were freshly encoded and fairly consistent. In this paper, we will cover the practical and technical challenges involved in preparing a large (900MiB) corpus of XML for ingest into an open-source archival information system (ArchivesSpace). This case study will give an overview of the project, discuss problem discovery and problem solving, and address the technical challenges, analysis, solutions, and decisions and provide information on the tools produced and lessons learned. The authors of this piece are Kate Bowers, Collections Services Archivist for Metadata, Systems, and Standards at the Harvard University Archive, and Dave Mayo, a Digital Library Software Engineer for Harvard's Library and Technology Services. Kate was heavily involved in both metadata analysis and later problem solving, while Dave was the sole full-time developer assigned to the migration project.
    Date
    31. 1.2017 13:29:56
  10. Clark, J.A.; Young, S.W.H.: Building a better book in the browser : using Semantic Web technologies and HTML5 (2015) 0.07
    0.06596371 = product of:
      0.13192742 = sum of:
        0.12042975 = weight(_text_:markup in 2116) [ClassicSimilarity], result of:
          0.12042975 = score(doc=2116,freq=2.0), product of:
            0.27638784 = queryWeight, product of:
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.042049456 = queryNorm
            0.43572736 = fieldWeight in 2116, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.046875 = fieldNorm(doc=2116)
        0.011497671 = product of:
          0.03449301 = sum of:
            0.03449301 = weight(_text_:29 in 2116) [ClassicSimilarity], result of:
              0.03449301 = score(doc=2116,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.23319192 = fieldWeight in 2116, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.046875 = fieldNorm(doc=2116)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    The library as place and service continues to be shaped by the legacy of the book. The book itself has evolved in recent years, with various technologies vying to become the next dominant book form. In this article, we discuss the design and development of our prototype software from Montana State University (MSU) Library for presenting books inside of web browsers. The article outlines the contextual background and technological potential for publishing traditional book content through the web using open standards. Our prototype demonstrates the application of HTML5, structured data with RDFa and Schema.org markup, linked data components using JSON-LD, and an API-driven data model. We examine how this open web model impacts discovery, reading analytics, eBook production, and machine-readability for libraries considering how to unite software development and publishing.
    Source
    Code4Lib journal. Issue 29(2015), [http://journal.code4lib.org/issues/issues/issue29]
  11. Miller, D.R.: XML: Libraries' strategic opportunity (2001) 0.06
    0.05613823 = product of:
      0.11227646 = sum of:
        0.10035812 = weight(_text_:markup in 1467) [ClassicSimilarity], result of:
          0.10035812 = score(doc=1467,freq=2.0), product of:
            0.27638784 = queryWeight, product of:
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.042049456 = queryNorm
            0.36310613 = fieldWeight in 1467, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1467)
        0.011918336 = product of:
          0.03575501 = sum of:
            0.03575501 = weight(_text_:language in 1467) [ClassicSimilarity], result of:
              0.03575501 = score(doc=1467,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.21673335 = fieldWeight in 1467, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1467)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    XML (eXtensible Markup Language) is fast gaining favor as the universal format for data and document exchange -- in effect becoming the lingua franca of the Information Age. Currently, "library information" is at a particular disadvantage on the rapidly evolving World Wide Web. Why? Despite libraries'explorations of web catalogs, scanning projects, digital data repositories, and creation of web pages galore, there remains a digital divide. The core of libraries' data troves are stored in proprietary formats of integrated library systems (ILS) and in the complex and arcane MARC formats -- both restricted chiefly to the province of technical services and systems librarians. Even they are hard-pressed to extract and integrate this wealth of data with resources from outside this rarefied environment. Segregation of library information underlies many difficulties: producing standard bibliographic citations from MARC data, automatically creating new materials lists (including new web resources) on a particular topic, exchanging data with our vendors, and even migrating from one ILS to another. Why do we continue to hobble our potential by embracing these self-imposed limitations? Most ILSs began in libraries, which soon recognized the pitfalls of do-it-yourself solutions. Thus, we wisely anticipated the necessity for standards. However, with the advent of the web, we soon found "our" collections and a flood of new resources appearing in digital format on opposite sides of the divide. If we do not act quickly to integrate library resources with mainstream web resources, we are in grave danger of becoming marginalized
  12. Adams, K.C.: Word wranglers : Automatic classification tools transform enterprise documents from "bags of words" into knowledge resources (2003) 0.06
    0.05613823 = product of:
      0.11227646 = sum of:
        0.10035812 = weight(_text_:markup in 1665) [ClassicSimilarity], result of:
          0.10035812 = score(doc=1665,freq=2.0), product of:
            0.27638784 = queryWeight, product of:
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.042049456 = queryNorm
            0.36310613 = fieldWeight in 1665, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1665)
        0.011918336 = product of:
          0.03575501 = sum of:
            0.03575501 = weight(_text_:language in 1665) [ClassicSimilarity], result of:
              0.03575501 = score(doc=1665,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.21673335 = fieldWeight in 1665, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1665)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    Taxonomies are an important part of any knowledge management (KM) system, and automatic classification software is emerging as a "killer app" for consumer and enterprise portals. A number of companies such as Inxight Software , Mohomine, Metacode, and others claim to interpret the semantic content of any textual document and automatically classify text on the fly. The promise that software could automatically produce a Yahoo-style directory is a siren call not many IT managers are able to resist. KM needs have grown more complex due to the increasing amount of digital information, the declining effectiveness of keyword searching, and heterogeneous document formats in corporate databases. This environment requires innovative KM tools, and automatic classification technology is an example of this new kind of software. These products can be divided into three categories according to their underlying technology - rules-based, catalog-by-example, and statistical clustering. Evolving trends in this market include framing classification as a cyborg (computer- and human-based) activity and the increasing use of extensible markup language (XML) and support vector machine (SVM) technology. In this article, we'll survey the rapidly changing automatic classification software market and examine the features and capabilities of leading classification products.
  13. Li, Z.: ¬A domain specific search engine with explicit document relations (2013) 0.06
    0.05613823 = product of:
      0.11227646 = sum of:
        0.10035812 = weight(_text_:markup in 1210) [ClassicSimilarity], result of:
          0.10035812 = score(doc=1210,freq=2.0), product of:
            0.27638784 = queryWeight, product of:
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.042049456 = queryNorm
            0.36310613 = fieldWeight in 1210, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.0390625 = fieldNorm(doc=1210)
        0.011918336 = product of:
          0.03575501 = sum of:
            0.03575501 = weight(_text_:language in 1210) [ClassicSimilarity], result of:
              0.03575501 = score(doc=1210,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.21673335 = fieldWeight in 1210, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1210)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    The current web consists of documents that are highly heterogeneous and hard for machines to understand. The Semantic Web is a progressive movement of the Word Wide Web, aiming at converting the current web of unstructured documents to the web of data. In the Semantic Web, web documents are annotated with metadata using standardized ontology language. These annotated documents are directly processable by machines and it highly improves their usability and usefulness. In Ericsson, similar problems occur. There are massive documents being created with well-defined structures. Though these documents are about domain specific knowledge and can have rich relations, they are currently managed by a traditional search engine, which ignores the rich domain specific information and presents few data to users. Motivated by the Semantic Web, we aim to find standard ways to process these documents, extract rich domain specific information and annotate these data to documents with formal markup languages. We propose this project to develop a domain specific search engine for processing different documents and building explicit relations for them. This research project consists of the three main focuses: examining different domain specific documents and finding ways to extract their metadata; integrating a text search engine with an ontology server; exploring novel ways to build relations for documents. We implement this system and demonstrate its functions. As a prototype, the system provides required features and will be extended in the future.
  14. Heery, R.; Wagner, H.: ¬A metadata registry for the Semantic Web (2002) 0.04
    0.039296757 = product of:
      0.078593515 = sum of:
        0.07025068 = weight(_text_:markup in 1210) [ClassicSimilarity], result of:
          0.07025068 = score(doc=1210,freq=2.0), product of:
            0.27638784 = queryWeight, product of:
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.042049456 = queryNorm
            0.2541743 = fieldWeight in 1210, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.02734375 = fieldNorm(doc=1210)
        0.008342835 = product of:
          0.025028504 = sum of:
            0.025028504 = weight(_text_:language in 1210) [ClassicSimilarity], result of:
              0.025028504 = score(doc=1210,freq=2.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.15171334 = fieldWeight in 1210, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.02734375 = fieldNorm(doc=1210)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Abstract
    * Agencies maintaining directories of data elements in a domain area in accordance with ISO/IEC 11179 (This standard specifies good practice for data element definition as well as the registration process. Example implementations are the National Health Information Knowledgebase hosted by the Australian Institute of Health and Welfare and the Environmental Data Registry hosted by the US Environmental Protection Agency.); * The xml.org directory of the Extended Markup Language (XML) document specifications facilitating re-use of Document Type Definition (DTD), hosted by the Organization for the Advancement of Structured Information Standards (OASIS); * The MetaForm database of Dublin Core usage and mappings maintained at the State and University Library in Goettingen; * The Semantic Web Agreement Group Dictionary, a database of terms for the Semantic Web that can be referred to by humans and software agents; * LEXML, a multi-lingual and multi-jurisdictional RDF Dictionary for the legal world; * The SCHEMAS registry maintained by the European Commission funded SCHEMAS project, which indexes several metadata element sets as well as a large number of activity reports describing metadata related activities and initiatives. Metadata registries essentially provide an index of terms. Given the distributed nature of the Web, there are a number of ways this can be accomplished. For example, the registry could link to terms and definitions in schemas published by implementers and stored locally by the schema maintainer. Alternatively, the registry might harvest various metadata schemas from their maintainers. Registries provide 'added value' to users by indexing schemas relevant to a particular 'domain' or 'community of use' and by simplifying the navigation of terms by enabling multiple schemas to be accessed from one view. An important benefit of this approach is an increase in the reuse of existing terms, rather than users having to reinvent them. Merging schemas to one view leads to harmonization between applications and helps avoid duplication of effort. Additionally, the establishment of registries to index terms actively being used in local implementations facilitates the metadata standards activity by providing implementation experience transferable to the standards-making process.
  15. Dietz, K.: en.wikipedia.org > 6 Mio. Artikel (2020) 0.03
    0.03261807 = product of:
      0.06523614 = sum of:
        0.05565474 = product of:
          0.16696422 = sum of:
            0.16696422 = weight(_text_:3a in 5669) [ClassicSimilarity], result of:
              0.16696422 = score(doc=5669,freq=2.0), product of:
                0.35649577 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.042049456 = queryNorm
                0.46834838 = fieldWeight in 5669, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5669)
          0.33333334 = coord(1/3)
        0.009581393 = product of:
          0.028744178 = sum of:
            0.028744178 = weight(_text_:29 in 5669) [ClassicSimilarity], result of:
              0.028744178 = score(doc=5669,freq=2.0), product of:
                0.14791684 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.042049456 = queryNorm
                0.19432661 = fieldWeight in 5669, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5669)
          0.33333334 = coord(1/3)
      0.5 = coord(2/4)
    
    Content
    "Die Englischsprachige Wikipedia verfügt jetzt über mehr als 6 Millionen Artikel. An zweiter Stelle kommt die deutschsprachige Wikipedia mit 2.3 Millionen Artikeln, an dritter Stelle steht die französischsprachige Wikipedia mit 2.1 Millionen Artikeln (via Researchbuzz: Firehose <https://rbfirehose.com/2020/01/24/techcrunch-wikipedia-now-has-more-than-6-million-articles-in-english/> und Techcrunch <https://techcrunch.com/2020/01/23/wikipedia-english-six-million-articles/?utm_source=feedburner&utm_medium=feed&utm_campaign=Feed%3A+Techcrunch+%28TechCrunch%29&guccounter=1&guce_referrer=aHR0cHM6Ly9yYmZpcmVob3NlLmNvbS8yMDIwLzAxLzI0L3RlY2hjcnVuY2gtd2lraXBlZGlhLW5vdy1oYXMtbW9yZS10aGFuLTYtbWlsbGlvbi1hcnRpY2xlcy1pbi1lbmdsaXNoLw&guce_referrer_sig=AQAAAK0zHfjdDZ_spFZBF_z-zDjtL5iWvuKDumFTzm4HvQzkUfE2pLXQzGS6FGB_y-VISdMEsUSvkNsg2U_NWQ4lwWSvOo3jvXo1I3GtgHpP8exukVxYAnn5mJspqX50VHIWFADHhs5AerkRn3hMRtf_R3F1qmEbo8EROZXp328HMC-o>). 250120 via digithek ch = #fineBlog s.a.: Angesichts der Veröffentlichung des 6-millionsten Artikels vergangene Woche in der englischsprachigen Wikipedia hat die Community-Zeitungsseite "Wikipedia Signpost" ein Moratorium bei der Veröffentlichung von Unternehmensartikeln gefordert. Das sei kein Vorwurf gegen die Wikimedia Foundation, aber die derzeitigen Maßnahmen, um die Enzyklopädie gegen missbräuchliches undeklariertes Paid Editing zu schützen, funktionierten ganz klar nicht. *"Da die ehrenamtlichen Autoren derzeit von Werbung in Gestalt von Wikipedia-Artikeln überwältigt werden, und da die WMF nicht in der Lage zu sein scheint, dem irgendetwas entgegenzusetzen, wäre der einzige gangbare Weg für die Autoren, fürs erste die Neuanlage von Artikeln über Unternehmen zu untersagen"*, schreibt der Benutzer Smallbones in seinem Editorial <https://en.wikipedia.org/wiki/Wikipedia:Wikipedia_Signpost/2020-01-27/From_the_editor> zur heutigen Ausgabe."
  16. Dachwitz, I.: ¬Das sind 650.000 Kategorien, in die uns die Online-Werbeindustrie einsortiert : Microsofts Datenmarktplatz Xandr (2023) 0.03
    0.028385563 = product of:
      0.11354225 = sum of:
        0.11354225 = weight(_text_:markup in 982) [ClassicSimilarity], result of:
          0.11354225 = score(doc=982,freq=4.0), product of:
            0.27638784 = queryWeight, product of:
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.042049456 = queryNorm
            0.4108077 = fieldWeight in 982, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              6.572923 = idf(docFreq=167, maxDocs=44218)
              0.03125 = fieldNorm(doc=982)
      0.25 = coord(1/4)
    
    Content
    "Was auch immer wir im Internet tun, wird aufgezeichnet und ausgewertet, um uns zielgerichtet Werbung anzuzeigen. Das ist eine Realität, an die viele Menschen sich inzwischen gewöhnt haben - im Gegenzug sind schließlich viele Internetangebote kostenlos. Wo genau unsere Daten landen, wenn wir Websites aufrufen oder Apps nutzen, das können die wenigsten nachvollziehen. Auch daran haben wir uns gewöhnt. Die Wege des Targeted Advertising sind unergründlich. Die Werbeindustrie tut viel dafür, damit das so bleibt: Die Netzwerke der Datensammler sind selbst für Branchenkenner:innen kaum zu überschauen. Jetzt präsentieren netzpolitik.org und das US-Medium The Markup einen einmaligen Einblick in das Geschäft mit unseren Daten. Wir haben die Angebotsliste von Xandr ausgewertet, einem der größten Datenmarktplätze der Werbewelt. Sie enthält mehr als 650.000 unterschiedliche Kategorien, in die die Industrie Menschen einsortiert, um sie mit gezielter Werbung erreichen zu können.
    Umfang und Detailtiefe dieser Datensammlung sind erschreckend. Es gibt kaum eine menschliche Eigenschaft, die Werbetreibende nicht für Werbung ausnutzen wollen. Sie wollen Menschen aus Dänemark erreichen, die einen Toyota gekauft haben? Kein Problem. Sie wollen Menschen erreichen, die gerade finanzielle Probleme haben? Oder keine Krankenversicherung? Kein Problem. Minderjährige? Schwangere? Homosexuelle? Depressive? Politiker:innen? Alles kein Problem. "Diese Liste ist das gewaltigste Dokument über den globalen Datenhandel, das ich je gesehen habe", sagt der Wiener Tracking-Forscher Wolfie Christl. Er hat die Datei aufgestöbert und mit netzpolitik.org sowie The Markup geteilt. Das US-Medium berichtet heute unter anderem über die zahlreichen sensiblen Daten und macht sie mit einem interaktiven Tool einfach durchsuchbar. Xandr hat auf mehrere Presseanfragen nicht reagiert. Die Liste ist auf Mai 2021 datiert, sie stand bis zu unserer Anfrage auf einer Dokumentationsseite von Xandr offen im Netz. Heute ist sie nicht mehr erreichbar, aber beim Internet Archive gibt es eine archivierte Version der Seite und der Datei [23 MB]. Laut von uns befragten Jurist:innen zeige die Liste, dass das derzeitige Werbegeschäft strukturell unvereinbar mit Datenschutzanforderungen ist."
  17. Kleineberg, M.: Context analysis and context indexing : formal pragmatics in knowledge organization (2014) 0.03
    0.02782737 = product of:
      0.11130948 = sum of:
        0.11130948 = product of:
          0.33392844 = sum of:
            0.33392844 = weight(_text_:3a in 1826) [ClassicSimilarity], result of:
              0.33392844 = score(doc=1826,freq=2.0), product of:
                0.35649577 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.042049456 = queryNorm
                0.93669677 = fieldWeight in 1826, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1826)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Source
    http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=5&ved=0CDQQFjAE&url=http%3A%2F%2Fdigbib.ubka.uni-karlsruhe.de%2Fvolltexte%2Fdocuments%2F3131107&ei=HzFWVYvGMsiNsgGTyoFI&usg=AFQjCNE2FHUeR9oQTQlNC4TPedv4Mo3DaQ&sig2=Rlzpr7a3BLZZkqZCXXN_IA&bvm=bv.93564037,d.bGg&cad=rja
  18. Panzer, M.: Designing identifiers for the DDC (2007) 0.03
    0.025858585 = product of:
      0.10343434 = sum of:
        0.10343434 = sum of:
          0.04797038 = weight(_text_:language in 1752) [ClassicSimilarity], result of:
            0.04797038 = score(doc=1752,freq=10.0), product of:
              0.16497234 = queryWeight, product of:
                3.9232929 = idf(docFreq=2376, maxDocs=44218)
                0.042049456 = queryNorm
              0.29077834 = fieldWeight in 1752, product of:
                3.1622777 = tf(freq=10.0), with freq of:
                  10.0 = termFreq=10.0
                3.9232929 = idf(docFreq=2376, maxDocs=44218)
                0.0234375 = fieldNorm(doc=1752)
          0.017246505 = weight(_text_:29 in 1752) [ClassicSimilarity], result of:
            0.017246505 = score(doc=1752,freq=2.0), product of:
              0.14791684 = queryWeight, product of:
                3.5176873 = idf(docFreq=3565, maxDocs=44218)
                0.042049456 = queryNorm
              0.11659596 = fieldWeight in 1752, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5176873 = idf(docFreq=3565, maxDocs=44218)
                0.0234375 = fieldNorm(doc=1752)
          0.03821745 = weight(_text_:22 in 1752) [ClassicSimilarity], result of:
            0.03821745 = score(doc=1752,freq=10.0), product of:
              0.14725003 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.042049456 = queryNorm
              0.2595412 = fieldWeight in 1752, product of:
                3.1622777 = tf(freq=10.0), with freq of:
                  10.0 = termFreq=10.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0234375 = fieldNorm(doc=1752)
      0.25 = coord(1/4)
    
    Content
    Some examples of identifiers for concepts follow: <http://dewey.info/concept/338.4/en/edn/22/> This identifier is used to retrieve or identify the 338.4 concept in the English-language version of Edition 22. <http://dewey.info/concept/338.4/de/edn/22/> This identifier is used to retrieve or identify the 338.4 concept in the German-language version of Edition 22. <http://dewey.info/concept/333.7-333.9/> This identifier is used to retrieve or identify the 333.7-333.9 concept across all editions and language versions. <http://dewey.info/concept/333.7-333.9/about.skos> This identifier is used to retrieve a SKOS representation of the 333.7-333.9 concept (using the "resource" element). There are several open issues at this preliminary stage of development: Use cases: URIs need to represent the range of statements or questions that could be submitted to a Dewey web service. Therefore, it seems that some general questions have to be answered first: What information does an agent have when coming to a Dewey web service? What kind of questions will such an agent ask? Placement of the {locale} component: It is still an open question if the {locale} component should be placed after the {version} component instead (<http://dewey.info/concept/338.4/edn/22/en>) to emphasize that the most important instantiation of a Dewey class is its edition, not its language version. From a services point of view, however, it could make more sense to keep the current arrangement, because users are more likely to come to the service with a present understanding of the language version they are seeking without knowing the specifics of a certain edition in which they are trying to find topics. Identification of other Dewey entities: The goal is to create a locator that does not answer all, but a lot of questions that could be asked about the DDC. Which entities are missing but should be surfaced for services or user agents? How will those services or agents interact with them? Should some entities be rendered in a different way as presented? For example, (how) should the DDC Summaries be retrievable? Would it be necessary to make the DDC Manual accessible through this identifier structure?"
    Date
    21. 3.2008 19:29:28
  19. OWL Web Ontology Language Test Cases (2004) 0.02
    0.024110693 = product of:
      0.096442774 = sum of:
        0.096442774 = product of:
          0.14466415 = sum of:
            0.09908719 = weight(_text_:language in 4685) [ClassicSimilarity], result of:
              0.09908719 = score(doc=4685,freq=6.0), product of:
                0.16497234 = queryWeight, product of:
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.042049456 = queryNorm
                0.60062915 = fieldWeight in 4685, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  3.9232929 = idf(docFreq=2376, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4685)
            0.045576964 = weight(_text_:22 in 4685) [ClassicSimilarity], result of:
              0.045576964 = score(doc=4685,freq=2.0), product of:
                0.14725003 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.042049456 = queryNorm
                0.30952093 = fieldWeight in 4685, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4685)
          0.6666667 = coord(2/3)
      0.25 = coord(1/4)
    
    Abstract
    This document contains and presents test cases for the Web Ontology Language (OWL) approved by the Web Ontology Working Group. Many of the test cases illustrate the correct usage of the Web Ontology Language (OWL), and the formal meaning of its constructs. Other test cases illustrate the resolution of issues considered by the Working Group. Conformance for OWL documents and OWL document checkers is specified.
    Date
    14. 8.2011 13:33:22
  20. Popper, K.R.: Three worlds : the Tanner lecture on human values. Deliverd at the University of Michigan, April 7, 1978 (1978) 0.02
    0.022261895 = product of:
      0.08904758 = sum of:
        0.08904758 = product of:
          0.26714274 = sum of:
            0.26714274 = weight(_text_:3a in 230) [ClassicSimilarity], result of:
              0.26714274 = score(doc=230,freq=2.0), product of:
                0.35649577 = queryWeight, product of:
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.042049456 = queryNorm
                0.7493574 = fieldWeight in 230, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  8.478011 = idf(docFreq=24, maxDocs=44218)
                  0.0625 = fieldNorm(doc=230)
          0.33333334 = coord(1/3)
      0.25 = coord(1/4)
    
    Source
    https%3A%2F%2Ftannerlectures.utah.edu%2F_documents%2Fa-to-z%2Fp%2Fpopper80.pdf&usg=AOvVaw3f4QRTEH-OEBmoYr2J_c7H

Years

Languages

  • e 267
  • d 167
  • a 3
  • el 3
  • es 1
  • i 1
  • nl 1
  • More… Less…

Types

  • a 205
  • i 20
  • n 10
  • p 8
  • m 7
  • r 6
  • b 5
  • x 5
  • s 4
  • More… Less…

Themes