Search (1127 results, page 3 of 57)

  • × type_ss:"el"
  1. Denton, W.: Putting facets on the Web : an annotated bibliography (2003) 0.02
    0.017455934 = product of:
      0.061095767 = sum of:
        0.016528286 = weight(_text_:wide in 2467) [ClassicSimilarity], result of:
          0.016528286 = score(doc=2467,freq=2.0), product of:
            0.13505316 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.030480823 = queryNorm
            0.122383565 = fieldWeight in 2467, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.01953125 = fieldNorm(doc=2467)
        0.03106221 = weight(_text_:web in 2467) [ClassicSimilarity], result of:
          0.03106221 = score(doc=2467,freq=24.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.3122631 = fieldWeight in 2467, product of:
              4.8989797 = tf(freq=24.0), with freq of:
                24.0 = termFreq=24.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.01953125 = fieldNorm(doc=2467)
        0.0058015957 = weight(_text_:information in 2467) [ClassicSimilarity], result of:
          0.0058015957 = score(doc=2467,freq=10.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.10842399 = fieldWeight in 2467, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.01953125 = fieldNorm(doc=2467)
        0.0077036773 = weight(_text_:retrieval in 2467) [ClassicSimilarity], result of:
          0.0077036773 = score(doc=2467,freq=2.0), product of:
            0.092201896 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.030480823 = queryNorm
            0.08355226 = fieldWeight in 2467, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.01953125 = fieldNorm(doc=2467)
      0.2857143 = coord(4/14)
    
    Abstract
    This is a classified, annotated bibliography about how to design faceted classification systems and make them usable on the World Wide Web. It is the first of three works I will be doing. The second, based on the material here and elsewhere, will discuss how to actually make the faceted system and put it online. The third will be a report of how I did just that, what worked, what didn't, and what I learned. Almost every article or book listed here begins with an explanation of what a faceted classification system is, so I won't (but see Steckel in Background below if you don't already know). They all agree that faceted systems are very appropriate for the web. Even pre-web articles (such as Duncan's in Background, below) assert that hypertext and facets will go together well. Combined, it is possible to take a set of documents and classify them or apply subject headings to describe what they are about, then build a navigational structure so that any user, no matter how he or she approaches the material, no matter what his or her goals, can move and search in a way that makes sense to them, but still get to the same useful results as someone else following a different path to the same goal. There is no one way that everyone will always use when looking for information. The more flexible the organization of the information, the more accommodating it is. Facets are more flexible for hypertext browsing than any enumerative or hierarchical system.
    Consider movie listings in newspapers. Most Canadian newspapers list movie showtimes in two large blocks, for the two major theatre chains. The listings are ordered by region (in large cities), then theatre, then movie, and finally by showtime. Anyone wondering where and when a particular movie is playing must scan the complete listings. Determining what movies are playing in the next half hour is very difficult. When movie listings went onto the web, most sites used a simple faceted organization, always with movie name and theatre, and perhaps with region or neighbourhood (thankfully, theatre chains were left out). They make it easy to pick a theatre and see what movies are playing there, or to pick a movie and see what theatres are showing it. To complete the system, the sites should allow users to browse by neighbourhood and showtime, and to order the results in any way they desired. Thus could people easily find answers to such questions as, "Where is the new James Bond movie playing?" "What's showing at the Roxy tonight?" "I'm going to be out in in Little Finland this afternoon with three hours to kill starting at 2 ... is anything interesting playing?" A hypertext, faceted classification system makes more useful information more easily available to the user. Reading the books and articles below in chronological order will show a certain progression: suggestions that faceting and hypertext might work well, confidence that facets would work well if only someone would make such a system, and finally the beginning of serious work on actually designing, building, and testing faceted web sites. There is a solid basis of how to make faceted classifications (see Vickery in Recommended), but their application online is just starting. Work on XFML (see Van Dijck's work in Recommended) the Exchangeable Faceted Metadata Language, will make this easier. If it follows previous patterns, parts of the Internet community will embrace the idea and make open source software available for others to reuse. It will be particularly beneficial if professionals in both information studies and computer science can work together to build working systems, standards, and code. Each can benefit from the other's expertise in what can be a very complicated and technical area. One particularly nice thing about this area of research is that people interested in combining facets and the web often have web sites where they post their writings.
    This bibliography is not meant to be exhaustive, but unfortunately it is not as complete as I wanted. Some books and articles are not be included, but they may be used in my future work. (These include two books and one article by B.C. Vickery: Faceted Classification Schemes (New Brunswick, NJ: Rutgers, 1966), Classification and Indexing in Science, 3rd ed. (London: Butterworths, 1975), and "Knowledge Representation: A Brief Review" (Journal of Documentation 42 no. 3 (September 1986): 145-159; and A.C. Foskett's "The Future of Faceted Classification" in The Future of Classification, edited by Rita Marcella and Arthur Maltby (Aldershot, England: Gower, 2000): 69-80). Nevertheless, I hope this bibliography will be useful for those both new to or familiar with faceted hypertext systems. Some very basic resources are listed, as well as some very advanced ones. Some example web sites are mentioned, but there is no detailed technical discussion of any software. The user interface to any web site is extremely important, and this is briefly mentioned in two or three places (for example the discussion of lawforwa.org (see Example Web Sites)). The larger question of how to display information graphically and with hypertext is outside the scope of this bibliography. There are five sections: Recommended, Background, Not Relevant, Example Web Sites, and Mailing Lists. Background material is either introductory, advanced, or of peripheral interest, and can be read after the Recommended resources if the reader wants to know more. The Not Relevant category contains articles that may appear in bibliographies but are not relevant for my purposes.
    Theme
    Klassifikationssysteme im Online-Retrieval
  2. Brake, M.: ¬Das Semantische Web : Eine Vision in der Halbzeit (2005) 0.02
    0.017203834 = product of:
      0.12042683 = sum of:
        0.043487098 = weight(_text_:web in 4764) [ClassicSimilarity], result of:
          0.043487098 = score(doc=4764,freq=6.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.43716836 = fieldWeight in 4764, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4764)
        0.07693974 = weight(_text_:wien in 4764) [ClassicSimilarity], result of:
          0.07693974 = score(doc=4764,freq=2.0), product of:
            0.17413543 = queryWeight, product of:
              5.7129507 = idf(docFreq=396, maxDocs=44218)
              0.030480823 = queryNorm
            0.4418385 = fieldWeight in 4764, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.7129507 = idf(docFreq=396, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4764)
      0.14285715 = coord(2/14)
    
    Abstract
    Mit der Veröffentlichung der ersten Website 1989 am CERN gab Timothy J. Berners-Lee den Startschuss für das WorldWideWeb, diesem ständig expandierenden Konvolut aus digitalisierten Daten, vernetzten Maschinen und Menschen. Mittlerweile nimmt das Netz für sich in Anspruch, das Weltwissen zu repräsentieren und hat in seiner Gesamtheit längst die klassische Vorstellung von einer Enzyklopädie abgelöst und übertroffen. Andererseits ist das Netz trotz geballter Wissensansammlung und Rechenpower immer noch dumm. Seit fast einer Dekade ist nun das "verständige" Internet in der Diskussion, die Semantic Web Days in München und die Semantics in Wien zogen eine Zwischenbilanz und zeigten ein mittlerweile weites Spektrum von Praxisanwendungen.
    Theme
    Semantic Web
  3. Bohne-Lang, A.: Semantische Metadaten für den Webauftritt einer Bibliothek (2016) 0.02
    0.017080726 = product of:
      0.07971005 = sum of:
        0.02536219 = weight(_text_:web in 3337) [ClassicSimilarity], result of:
          0.02536219 = score(doc=3337,freq=4.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.25496176 = fieldWeight in 3337, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3337)
        0.049158752 = weight(_text_:bibliothek in 3337) [ClassicSimilarity], result of:
          0.049158752 = score(doc=3337,freq=6.0), product of:
            0.12513994 = queryWeight, product of:
              4.1055303 = idf(docFreq=1980, maxDocs=44218)
              0.030480823 = queryNorm
            0.39283025 = fieldWeight in 3337, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.1055303 = idf(docFreq=1980, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3337)
        0.005189105 = weight(_text_:information in 3337) [ClassicSimilarity], result of:
          0.005189105 = score(doc=3337,freq=2.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.09697737 = fieldWeight in 3337, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=3337)
      0.21428572 = coord(3/14)
    
    Abstract
    Das Semantic Web ist schon seit über 10 Jahren viel beachtet und hat mit der Verfügbarkeit von Resource Description Framework (RDF) und den entsprechenden Ontologien einen großen Sprung in die Praxis gemacht. Vertreter kleiner Bibliotheken und Bibliothekare mit geringer Technik-Affinität stehen aber im Alltag vor großen Hürden, z.B. bei der Frage, wie man diese Technik konkret in den eigenen Webauftritt einbinden kann: man kommt sich vor wie Don Quijote, der versucht die Windmühlen zu bezwingen. RDF mit seinen Ontologien ist fast unverständlich komplex für Nicht-Informatiker und somit für den praktischen Einsatz auf Bibliotheksseiten in der Breite nicht direkt zu gebrauchen. Mit Schema.org wurde ursprünglich von den drei größten Suchmaschinen der Welt Google, Bing und Yahoo eine einfach und effektive semantische Beschreibung von Entitäten entwickelt. Aktuell wird Schema.org durch Google, Microsoft, Yahoo und Yandex weiter gesponsert und von vielen weiteren Suchmaschinen verstanden. Vor diesem Hintergrund hat die Bibliothek der Medizinischen Fakultät Mannheim auf ihrer Homepage (http://www.umm.uni-heidelberg.de/bibl/) verschiedene maschinenlesbare semantische Metadaten eingebettet. Sehr interessant und zukunftsweisend ist die neueste Entwicklung von Schema.org, bei der man eine 'Library' (https://schema.org/Library) mit Öffnungszeiten und vielem mehr modellieren kann. Ferner haben wir noch semantische Metadaten im Open Graph- und Dublin Core-Format eingebettet, um alte Standards und Facebook-konforme Informationen maschinenlesbar zur Verfügung zu stellen.
    Source
    GMS Medizin - Bibliothek - Information. 16(2016) Nr.3, 11 S. [http://www.egms.de/static/pdf/journals/mbi/2017-16/mbi000372.pdf]
    Theme
    Semantic Web
  4. Pepper, S.; Moore, G.; TopicMaps.Org Authoring Group: XML Topic Maps (XTM) 1.0 : TopicMaps.Org Specification (2001) 0.02
    0.01690901 = product of:
      0.07890872 = sum of:
        0.039667886 = weight(_text_:wide in 1623) [ClassicSimilarity], result of:
          0.039667886 = score(doc=1623,freq=2.0), product of:
            0.13505316 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.030480823 = queryNorm
            0.29372054 = fieldWeight in 1623, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046875 = fieldNorm(doc=1623)
        0.030434625 = weight(_text_:web in 1623) [ClassicSimilarity], result of:
          0.030434625 = score(doc=1623,freq=4.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.3059541 = fieldWeight in 1623, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=1623)
        0.0088062035 = weight(_text_:information in 1623) [ClassicSimilarity], result of:
          0.0088062035 = score(doc=1623,freq=4.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.16457605 = fieldWeight in 1623, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=1623)
      0.21428572 = coord(3/14)
    
    Abstract
    This specification provides a model and grammar for representing the structure of information resources used to define topics, and the associations (relationships) between topics. Names, resources, and relationships are said to be characteristics of abstract subjects, which are called topics. Topics have their characteristics within scopes: i.e. the limited contexts within which the names and resources are regarded as their name, resource, and relationship characteristics. One or more interrelated documents employing this grammar is called a topic map.TopicMaps.Org is an independent consortium of parties developing the applicability of the topic map paradigm [ISO13250] to the World Wide Web by leveraging the XML family of specifications. This specification describes version 1.0 of XML Topic Maps (XTM) 1.0 [XTM], an abstract model and XML grammar for interchanging Web-based topic maps, written by the members of the TopicMaps.Org Authoring Group. More information on XTM and TopicMaps.Org is available at http://www.topicmaps.org/about.html. All versions of the XTM Specification are permanently licensed to the public, as provided by the Charter of TopicMaps.Org.
  5. Metadata practices on the cutting edge (2004) 0.02
    0.016853835 = product of:
      0.078651235 = sum of:
        0.046279203 = weight(_text_:wide in 2335) [ClassicSimilarity], result of:
          0.046279203 = score(doc=2335,freq=2.0), product of:
            0.13505316 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.030480823 = queryNorm
            0.342674 = fieldWeight in 2335, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2335)
        0.025107287 = weight(_text_:web in 2335) [ClassicSimilarity], result of:
          0.025107287 = score(doc=2335,freq=2.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.25239927 = fieldWeight in 2335, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2335)
        0.0072647473 = weight(_text_:information in 2335) [ClassicSimilarity], result of:
          0.0072647473 = score(doc=2335,freq=2.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.13576832 = fieldWeight in 2335, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0546875 = fieldNorm(doc=2335)
      0.21428572 = coord(3/14)
    
    Abstract
    The PowerPoint presentations from this one-day workshop on emerging metadata practices are available at this web site. Topics include metadata quality, interoperability, linking metadata, metadata for image collections, RSS, MODS, METS, and MPEG-21. Contributors include representatives from OCLC, CrossRef, the Library of Congress, universities and the private sector. Given the wide range of presentations, if you're interested in metadata you can likely find something of interest here, but no single topic is explored in much depth, and you are sometimes left wondering what the speaker said about a particular slide if there are no accompanying notes.
    Imprint
    Washington, DC : National Information Standards Organization
  6. Weller, K.: Ontologien: Stand und Entwicklung der Semantik für WorldWideWeb (2009) 0.02
    0.016788604 = product of:
      0.07834682 = sum of:
        0.033056572 = weight(_text_:wide in 4425) [ClassicSimilarity], result of:
          0.033056572 = score(doc=4425,freq=2.0), product of:
            0.13505316 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.030480823 = queryNorm
            0.24476713 = fieldWeight in 4425, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4425)
        0.04010114 = weight(_text_:web in 4425) [ClassicSimilarity], result of:
          0.04010114 = score(doc=4425,freq=10.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.40312994 = fieldWeight in 4425, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4425)
        0.005189105 = weight(_text_:information in 4425) [ClassicSimilarity], result of:
          0.005189105 = score(doc=4425,freq=2.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.09697737 = fieldWeight in 4425, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4425)
      0.21428572 = coord(3/14)
    
    Abstract
    Die Idee zu einem semantischen Web wurde maßgeblich geprägt (wenn auch nicht initiiert) durch eine Veröffentlichung von Tim Berners Lee, James Hendler und Ora Lassila im Jahre 2001. Darin skizzieren die Autoren ihre Version von einem erweiterten und verbesserten World Wide Web: Daten sollen so aufbereitet werden, dass nicht nur Menschen diese lesen können, sondern dass auch Computer in die Lage versetzt werden, diese zu verarbeiten und sinnvoll zu kombinieren. Sie beschreiben ein Szenario, in dem "Web agents" dem Nutzer bei der Durchführung komplexer Suchanfragen helfen, wie beispielsweise "finde einen Arzt, der eine bestimmte Behandlung anbietet, dessen Praxis in der Nähe meiner Wohnung liegt und dessen Öffnungszeiten mit meinem Terminkalender zusammenpassen". Die große Herausforderung liegt hierbei darin, dass Informationen, die über mehrere Webseiten verteilt sind, gesammelt und zu einer sinnvollen Antwort kombiniert werden müssen. Man spricht dabei vom Problem der Informationsintegration (Information Integration). Diese Vision der weltweiten Datenintegration in einem Semantic Web wurde seither vielfach diskutiert, erweitert und modifiziert, an der technischen Realisation arbeitet eine Vielzahl verschiedener Forschungseinrichtungen. Einigkeit besteht dahingehend, dass eine solche Idee nur mit der Hilfe neuer bedeutungstragender Metadaten verwirklicht werden kann. Benötigt werden also neue Ansätze zur Indexierung von Web Inhalten, die eine Suche über Wortbedeutungen und nicht über bloße Zeichenketten ermöglichen können. So soll z.B. erkannt werden, dass es sich bei "Heinrich Heine" um den Namen einer Person handelt und bei "Düsseldorf" um den Namen einer Stadt. Darüber hinaus sollen auch Verbindungen zwischen einzelnen Informationseinheiten festgehalten werden, beispielsweise dass Heinrich Heine in Düsseldorf wohnte. Wenn solche semantischen Relationen konsequent eingesetzt werden, können sie in vielen Fällen ausgenutzt werden, um neue Schlussfolgerungen zu ziehen.
  7. Tomassen, S.L.: Research on ontology-driven information retrieval (2006 (?)) 0.02
    0.016756635 = product of:
      0.07819763 = sum of:
        0.030434625 = weight(_text_:web in 4328) [ClassicSimilarity], result of:
          0.030434625 = score(doc=4328,freq=4.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.3059541 = fieldWeight in 4328, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=4328)
        0.010785352 = weight(_text_:information in 4328) [ClassicSimilarity], result of:
          0.010785352 = score(doc=4328,freq=6.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.20156369 = fieldWeight in 4328, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=4328)
        0.03697765 = weight(_text_:retrieval in 4328) [ClassicSimilarity], result of:
          0.03697765 = score(doc=4328,freq=8.0), product of:
            0.092201896 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.030480823 = queryNorm
            0.40105087 = fieldWeight in 4328, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=4328)
      0.21428572 = coord(3/14)
    
    Abstract
    An increasing number of recent information retrieval systems make use of ontologies to help the users clarify their information needs and come up with semantic representations of documents. A particular concern here is the integration of these semantic approaches with traditional search technology. The research presented in this paper examines how ontologies can be efficiently applied to large-scale search systems for the web. We describe how these systems can be enriched with adapted ontologies to provide both an in-depth understanding of the user's needs as well as an easy integration with standard vector-space retrieval systems. The ontology concepts are adapted to the domain terminology by computing a feature vector for each concept. Later, the feature vectors are used to enrich a provided query. The whole retrieval system is under development as part of a larger Semantic Web standardization project for the Norwegian oil & gas sector.
  8. Zhang, J.; Mostafa, J.; Tripathy, H.: Information retrieval by semantic analysis and visualization of the concept space of D-Lib® magazine (2002) 0.02
    0.01665492 = product of:
      0.058292218 = sum of:
        0.016528286 = weight(_text_:wide in 1211) [ClassicSimilarity], result of:
          0.016528286 = score(doc=1211,freq=2.0), product of:
            0.13505316 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.030480823 = queryNorm
            0.122383565 = fieldWeight in 1211, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.01953125 = fieldNorm(doc=1211)
        0.008966887 = weight(_text_:web in 1211) [ClassicSimilarity], result of:
          0.008966887 = score(doc=1211,freq=2.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.09014259 = fieldWeight in 1211, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.01953125 = fieldNorm(doc=1211)
        0.011007754 = weight(_text_:information in 1211) [ClassicSimilarity], result of:
          0.011007754 = score(doc=1211,freq=36.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.20572007 = fieldWeight in 1211, product of:
              6.0 = tf(freq=36.0), with freq of:
                36.0 = termFreq=36.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.01953125 = fieldNorm(doc=1211)
        0.021789288 = weight(_text_:retrieval in 1211) [ClassicSimilarity], result of:
          0.021789288 = score(doc=1211,freq=16.0), product of:
            0.092201896 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.030480823 = queryNorm
            0.23632148 = fieldWeight in 1211, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.01953125 = fieldNorm(doc=1211)
      0.2857143 = coord(4/14)
    
    Abstract
    In this article we present a method for retrieving documents from a digital library through a visual interface based on automatically generated concepts. We used a vocabulary generation algorithm to generate a set of concepts for the digital library and a technique called the max-min distance technique to cluster them. Additionally, the concepts were visualized in a spring embedding graph layout to depict the semantic relationship among them. The resulting graph layout serves as an aid to users for retrieving documents. An online archive containing the contents of D-Lib Magazine from July 1995 to May 2002 was used to test the utility of an implemented retrieval and visualization system. We believe that the method developed and tested can be applied to many different domains to help users get a better understanding of online document collections and to minimize users' cognitive load during execution of search tasks. Over the past few years, the volume of information available through the World Wide Web has been expanding exponentially. Never has so much information been so readily available and shared among so many people. Unfortunately, the unstructured nature and huge volume of information accessible over networks have made it hard for users to sift through and find relevant information. To deal with this problem, information retrieval (IR) techniques have gained more intensive attention from both industrial and academic researchers. Numerous IR techniques have been developed to help deal with the information overload problem. These techniques concentrate on mathematical models and algorithms for retrieval. Popular IR models such as the Boolean model, the vector-space model, the probabilistic model and their variants are well established.
    From the user's perspective, however, it is still difficult to use current information retrieval systems. Users frequently have problems expressing their information needs and translating those needs into queries. This is partly due to the fact that information needs cannot be expressed appropriately in systems terms. It is not unusual for users to input search terms that are different from the index terms information systems use. Various methods have been proposed to help users choose search terms and articulate queries. One widely used approach is to incorporate into the information system a thesaurus-like component that represents both the important concepts in a particular subject area and the semantic relationships among those concepts. Unfortunately, the development and use of thesauri is not without its own problems. The thesaurus employed in a specific information system has often been developed for a general subject area and needs significant enhancement to be tailored to the information system where it is to be used. This thesaurus development process, if done manually, is both time consuming and labor intensive. Usage of a thesaurus in searching is complex and may raise barriers for the user. For illustration purposes, let us consider two scenarios of thesaurus usage. In the first scenario the user inputs a search term and the thesaurus then displays a matching set of related terms. Without an overview of the thesaurus - and without the ability to see the matching terms in the context of other terms - it may be difficult to assess the quality of the related terms in order to select the correct term. In the second scenario the user browses the whole thesaurus, which is organized as in an alphabetically ordered list. The problem with this approach is that the list may be long, and neither does it show users the global semantic relationship among all the listed terms.
    Nevertheless, because thesaurus use has shown to improve retrieval, for our method we integrate functions in the search interface that permit users to explore built-in search vocabularies to improve retrieval from digital libraries. Our method automatically generates the terms and their semantic relationships representing relevant topics covered in a digital library. We call these generated terms the "concepts", and the generated terms and their semantic relationships we call the "concept space". Additionally, we used a visualization technique to display the concept space and allow users to interact with this space. The automatically generated term set is considered to be more representative of subject area in a corpus than an "externally" imposed thesaurus, and our method has the potential of saving a significant amount of time and labor for those who have been manually creating thesauri as well. Information visualization is an emerging discipline and developed very quickly in the last decade. With growing volumes of documents and associated complexities, information visualization has become increasingly important. Researchers have found information visualization to be an effective way to use and understand information while minimizing a user's cognitive load. Our work was based on an algorithmic approach of concept discovery and association. Concepts are discovered using an algorithm based on an automated thesaurus generation procedure. Subsequently, similarities among terms are computed using the cosine measure, and the associations among terms are established using a method known as max-min distance clustering. The concept space is then visualized in a spring embedding graph, which roughly shows the semantic relationships among concepts in a 2-D visual representation. The semantic space of the visualization is used as a medium for users to retrieve the desired documents. In the remainder of this article, we present our algorithmic approach of concept generation and clustering, followed by description of the visualization technique and interactive interface. The paper ends with key conclusions and discussions on future work.
    Theme
    Semantisches Umfeld in Indexierung u. Retrieval
  9. Dolin, R.; Agrawal, D.; El Abbadi, A.; Pearlman, J.: Using automated classification for summarizing and selecting heterogeneous information sources (1998) 0.02
    0.016418636 = product of:
      0.05746522 = sum of:
        0.019833943 = weight(_text_:wide in 1253) [ClassicSimilarity], result of:
          0.019833943 = score(doc=1253,freq=2.0), product of:
            0.13505316 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.030480823 = queryNorm
            0.14686027 = fieldWeight in 1253, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1253)
        0.015217313 = weight(_text_:web in 1253) [ClassicSimilarity], result of:
          0.015217313 = score(doc=1253,freq=4.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.15297705 = fieldWeight in 1253, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1253)
        0.009340389 = weight(_text_:information in 1253) [ClassicSimilarity], result of:
          0.009340389 = score(doc=1253,freq=18.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.17455927 = fieldWeight in 1253, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1253)
        0.013073574 = weight(_text_:retrieval in 1253) [ClassicSimilarity], result of:
          0.013073574 = score(doc=1253,freq=4.0), product of:
            0.092201896 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.030480823 = queryNorm
            0.1417929 = fieldWeight in 1253, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.0234375 = fieldNorm(doc=1253)
      0.2857143 = coord(4/14)
    
    Abstract
    Information retrieval over the Internet increasingly requires the filtering of thousands of heterogeneous information sources. Important sources of information include not only traditional databases with structured data and queries, but also increasing numbers of non-traditional, semi- or unstructured collections such as Web sites, FTP archives, etc. As the number and variability of sources increases, new ways of automatically summarizing, discovering, and selecting collections relevant to a user's query are needed. One such method involves the use of classification schemes, such as the Library of Congress Classification (LCC), within which a collection may be represented based on its content, irrespective of the structure of the actual data or documents. For such a system to be useful in a large-scale distributed environment, it must be easy to use for both collection managers and users. As a result, it must be possible to classify documents automatically within a classification scheme. Furthermore, there must be a straightforward and intuitive interface with which the user may use the scheme to assist in information retrieval (IR). Our work with the Alexandria Digital Library (ADL) Project focuses on geo-referenced information, whether text, maps, aerial photographs, or satellite images. As a result, we have emphasized techniques which work with both text and non-text, such as combined textual and graphical queries, multi-dimensional indexing, and IR methods which are not solely dependent on words or phrases. Part of this work involves locating relevant online sources of information. In particular, we have designed and are currently testing aspects of an architecture, Pharos, which we believe will scale up to 1.000.000 heterogeneous sources. Pharos accommodates heterogeneity in content and format, both among multiple sources as well as within a single source. That is, we consider sources to include Web sites, FTP archives, newsgroups, and full digital libraries; all of these systems can include a wide variety of content and multimedia data formats. Pharos is based on the use of hierarchical classification schemes. These include not only well-known 'subject' (or 'concept') based schemes such as the Dewey Decimal System and the LCC, but also, for example, geographic classifications, which might be constructed as layers of smaller and smaller hierarchical longitude/latitude boxes. Pharos is designed to work with sophisticated queries which utilize subjects, geographical locations, temporal specifications, and other types of information domains. The Pharos architecture requires that hierarchically structured collection metadata be extracted so that it can be partitioned in such a way as to greatly enhance scalability. Automated classification is important to Pharos because it allows information sources to extract the requisite collection metadata automatically that must be distributed.
  10. Artemenko, O.; Shramko, M.: Entwicklung eines Werkzeugs zur Sprachidentifikation in mono- und multilingualen Texten (2005) 0.02
    0.015803045 = product of:
      0.055310655 = sum of:
        0.023139602 = weight(_text_:wide in 572) [ClassicSimilarity], result of:
          0.023139602 = score(doc=572,freq=2.0), product of:
            0.13505316 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.030480823 = queryNorm
            0.171337 = fieldWeight in 572, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.02734375 = fieldNorm(doc=572)
        0.017753532 = weight(_text_:web in 572) [ClassicSimilarity], result of:
          0.017753532 = score(doc=572,freq=4.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.17847323 = fieldWeight in 572, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.02734375 = fieldNorm(doc=572)
        0.0036323736 = weight(_text_:information in 572) [ClassicSimilarity], result of:
          0.0036323736 = score(doc=572,freq=2.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.06788416 = fieldWeight in 572, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02734375 = fieldNorm(doc=572)
        0.010785148 = weight(_text_:retrieval in 572) [ClassicSimilarity], result of:
          0.010785148 = score(doc=572,freq=2.0), product of:
            0.092201896 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.030480823 = queryNorm
            0.11697317 = fieldWeight in 572, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.02734375 = fieldNorm(doc=572)
      0.2857143 = coord(4/14)
    
    Abstract
    Mit der Verbreitung des Internets vermehrt sich die Menge der im World Wide Web verfügbaren Dokumente. Die Gewährleistung eines effizienten Zugangs zu gewünschten Informationen für die Internetbenutzer wird zu einer großen Herausforderung an die moderne Informationsgesellschaft. Eine Vielzahl von Werkzeugen wird bereits eingesetzt, um den Nutzern die Orientierung in der wachsenden Informationsflut zu erleichtern. Allerdings stellt die enorme Menge an unstrukturierten und verteilten Informationen nicht die einzige Schwierigkeit dar, die bei der Entwicklung von Werkzeugen dieser Art zu bewältigen ist. Die zunehmende Vielsprachigkeit von Web-Inhalten resultiert in dem Bedarf an Sprachidentifikations-Software, die Sprache/en von elektronischen Dokumenten zwecks gezielter Weiterverarbeitung identifiziert. Solche Sprachidentifizierer können beispielsweise effektiv im Bereich des Multilingualen Information Retrieval eingesetzt werden, da auf den Sprachidentifikationsergebnissen Prozesse der automatischen Indexbildung wie Stemming, Stoppwörterextraktion etc. aufbauen. In der vorliegenden Arbeit wird das neue System "LangIdent" zur Sprachidentifikation von elektronischen Textdokumenten vorgestellt, das in erster Linie für Lehre und Forschung an der Universität Hildesheim verwendet werden soll. "LangIdent" enthält eine Auswahl von gängigen Algorithmen zu der monolingualen Sprachidentifikation, die durch den Benutzer interaktiv ausgewählt und eingestellt werden können. Zusätzlich wurde im System ein neuer Algorithmus implementiert, der die Identifikation von Sprachen, in denen ein multilinguales Dokument verfasst ist, ermöglicht. Die Identifikation beschränkt sich nicht nur auf eine Aufzählung von gefundenen Sprachen, vielmehr wird der Text in monolinguale Abschnitte aufgeteilt, jeweils mit der Angabe der identifizierten Sprache.
  11. Berners-Lee, T.: ¬The Father of the Web will give the Internet back to the people (2018) 0.02
    0.015778167 = product of:
      0.07363144 = sum of:
        0.037399244 = weight(_text_:wide in 4495) [ClassicSimilarity], result of:
          0.037399244 = score(doc=4495,freq=4.0), product of:
            0.13505316 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.030480823 = queryNorm
            0.2769224 = fieldWeight in 4495, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.03125 = fieldNorm(doc=4495)
        0.032080915 = weight(_text_:web in 4495) [ClassicSimilarity], result of:
          0.032080915 = score(doc=4495,freq=10.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.32250395 = fieldWeight in 4495, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=4495)
        0.004151284 = weight(_text_:information in 4495) [ClassicSimilarity], result of:
          0.004151284 = score(doc=4495,freq=2.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.0775819 = fieldWeight in 4495, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=4495)
      0.21428572 = coord(3/14)
    
    Content
    "This week, Berners-Lee will launch Inrupt ( https://www.password-online.de/?email_id=571&user_id=1045&urlpassed=aHR0cHM6Ly93d3cuaW5ydXB0LmNvbQ&controller=stats&action=analyse&wysija-page=1&wysijap=subscriptions ), a startup that he has been building, in stealth mode, for the past nine months. For years now, Berners-Lee and other internet activists have been dreaming of a digital utopia where individuals control their own data and the internet remains free and open. But for Berners-Lee, the time for dreaming is over. "We have to do it now," he says, displaying an intensity and urgency that is uncharacteristic for this soft-spoken academic. "It's a historical moment." If all goes as planned, Inrupt will be to Solid what Netscape once was for many first-time users of the web: an easy way in. . On his screen, there is a simple-looking web page with tabs across the top: Tim's to-do list, his calendar, chats, address book. He built this app-one of the first on Solid for his personal use. It is simple, spare. In fact, it's so plain that, at first glance, it's hard to see its significance. But to Berners-Lee, this is where the revolution begins. The app, using Solid's decentralized technology, allows Berners-Lee to access all of his data seamlessly-his calendar, his music library, videos, chat, research. It's like a mashup of Google Drive, Microsoft Outlook, Slack, Spotify, and WhatsApp. The difference here is that, on Solid, all the information is under his control. In: Exclusive: Tim Berners-Lee tells us his radical new plan to upend the World Wide Web ( https://www.password-online.de/?email_id=571&user_id=1045&urlpassed=aHR0cHM6Ly93d3cuZmFzdGNvbXBhbnkuY29tLzkwMjQzOTM2L2V4Y2x1c2l2ZS10aW0tYmVybmVycy1sZWUtdGVsbHMtdXMtaGlzLXJhZGljYWwtbmV3LXBsYW4tdG8tdXBlbmQtdGhlLXdvcmxkLXdpZGUtd2Vi&controller=stats&action=analyse&wysija-page=1&wysijap=subscriptions ), in: https://www.fastcompany.com/90243936/exclusive-tim-berners-lee-tells-us-his-radical-new-plan-to-upend-the-world-wide-web ( https://www.password-online.de/?email_id=571&user_id=1045&urlpassed=aHR0cHM6Ly93d3cuZmFzdGNvbXBhbnkuY29tLzkwMjQzOTM2L2V4Y2x1c2l2ZS10aW0tYmVybmVycy1sZWUtdGVsbHMtdXMtaGlzLXJhZGljYWwtbmV3LXBsYW4tdG8tdXBlbmQtdGhlLXdvcmxkLXdpZGUtd2Vi&controller=stats&action=analyse&wysija-page=1&wysijap=subscriptions)."
  12. Resource Description Framework (RDF) (2004) 0.02
    0.015754087 = product of:
      0.1102786 = sum of:
        0.052890517 = weight(_text_:wide in 3063) [ClassicSimilarity], result of:
          0.052890517 = score(doc=3063,freq=2.0), product of:
            0.13505316 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.030480823 = queryNorm
            0.3916274 = fieldWeight in 3063, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.0625 = fieldNorm(doc=3063)
        0.057388082 = weight(_text_:web in 3063) [ClassicSimilarity], result of:
          0.057388082 = score(doc=3063,freq=8.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.5769126 = fieldWeight in 3063, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0625 = fieldNorm(doc=3063)
      0.14285715 = coord(2/14)
    
    Abstract
    The Resource Description Framework (RDF) integrates a variety of applications from library catalogs and world-wide directories to syndication and aggregation of news, software, and content to personal collections of music, photos, and events using XML as an interchange syntax. The RDF specifications provide a lightweight ontology system to support the exchange of knowledge on the Web. The W3C Semantic Web Activity Statement explains W3C's plans for RDF, including the RDF Core WG, Web Ontology and the RDF Interest Group.
    Theme
    Semantic Web
  13. Patton, G.; Hengel-Dittrich, C.; O'Neill, E.T.; Tillett, B.B.: VIAF (Virtual International Authority File) : Linking Die Deutsche Bibliothek and Library of Congress Name Authority Files (2006) 0.02
    0.015488921 = product of:
      0.07228163 = sum of:
        0.017933775 = weight(_text_:web in 6105) [ClassicSimilarity], result of:
          0.017933775 = score(doc=6105,freq=2.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.18028519 = fieldWeight in 6105, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6105)
        0.049158752 = weight(_text_:bibliothek in 6105) [ClassicSimilarity], result of:
          0.049158752 = score(doc=6105,freq=6.0), product of:
            0.12513994 = queryWeight, product of:
              4.1055303 = idf(docFreq=1980, maxDocs=44218)
              0.030480823 = queryNorm
            0.39283025 = fieldWeight in 6105, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.1055303 = idf(docFreq=1980, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6105)
        0.005189105 = weight(_text_:information in 6105) [ClassicSimilarity], result of:
          0.005189105 = score(doc=6105,freq=2.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.09697737 = fieldWeight in 6105, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=6105)
      0.21428572 = coord(3/14)
    
    Abstract
    Die Deutsche Bibliothek, the Library of Congress, and OCLC Online Computer Library Center are jointly developing a virtual international authority file (VIAF) for personal names which links authority records from the world's national bibliographic agencies and will be made freely available on the Web. The goals of the project are to prove the viability of automatically linking authority records from different national authority files and to demonstrate its benefits. The authority and bibliographic files from the Library of Congress and Die Deutsche Bibliothek were used to create the initial VIAF which contains over six million names with over a half million links. A key aspect of the project was the development of automated name matching algorithms which use information from both authority records and the corresponding bibliographic records. The practicality of algorithmically linking the personal names between national authority files was demonstrated; seventy percent of the authority records for personal names common to both files were automatically linked with an error rate of less than one percent. The long-term goal of the VIAF project is to combine the authoritative names from many national libraries and other significant sources into a shared global authority service.
  14. Kara, S.: ¬An ontology-based retrieval system using semantic indexing (2012) 0.02
    0.015204006 = product of:
      0.07095203 = sum of:
        0.021520529 = weight(_text_:web in 3829) [ClassicSimilarity], result of:
          0.021520529 = score(doc=3829,freq=2.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.21634221 = fieldWeight in 3829, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=3829)
        0.012453852 = weight(_text_:information in 3829) [ClassicSimilarity], result of:
          0.012453852 = score(doc=3829,freq=8.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.23274569 = fieldWeight in 3829, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=3829)
        0.03697765 = weight(_text_:retrieval in 3829) [ClassicSimilarity], result of:
          0.03697765 = score(doc=3829,freq=8.0), product of:
            0.092201896 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.030480823 = queryNorm
            0.40105087 = fieldWeight in 3829, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.046875 = fieldNorm(doc=3829)
      0.21428572 = coord(3/14)
    
    Abstract
    In this thesis, we present an ontology-based information extraction and retrieval system and its application to soccer domain. In general, we deal with three issues in semantic search, namely, usability, scalability and retrieval performance. We propose a keyword-based semantic retrieval approach. The performance of the system is improved considerably using domain-specific information extraction, inference and rules. Scalability is achieved by adapting a semantic indexing approach. The system is implemented using the state-of-the-art technologies in SemanticWeb and its performance is evaluated against traditional systems as well as the query expansion methods. Furthermore, a detailed evaluation is provided to observe the performance gain due to domain-specific information extraction and inference. Finally, we show how we use semantic indexing to solve simple structural ambiguities.
    Source
    Information Systems. 37(2012) no. 4, S.294-305
    Theme
    Semantic Web
  15. Van de Sompel, H.; Hochstenbach, P.: Reference linking in a hybrid library environment : part 1: frameworks for linking (1999) 0.02
    0.015174966 = product of:
      0.07081651 = sum of:
        0.037399244 = weight(_text_:wide in 1244) [ClassicSimilarity], result of:
          0.037399244 = score(doc=1244,freq=4.0), product of:
            0.13505316 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.030480823 = queryNorm
            0.2769224 = fieldWeight in 1244, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.03125 = fieldNorm(doc=1244)
        0.02028975 = weight(_text_:web in 1244) [ClassicSimilarity], result of:
          0.02028975 = score(doc=1244,freq=4.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.2039694 = fieldWeight in 1244, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=1244)
        0.013127513 = weight(_text_:information in 1244) [ClassicSimilarity], result of:
          0.013127513 = score(doc=1244,freq=20.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.2453355 = fieldWeight in 1244, product of:
              4.472136 = tf(freq=20.0), with freq of:
                20.0 = termFreq=20.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=1244)
      0.21428572 = coord(3/14)
    
    Abstract
    The creation of services linking related information entities is an area that is attracting an ever increasing interest in the ongoing development of the World Wide Web in general, and of research-related information systems in particular. Currently, both practice and theory point at linking services as being a major domain for innovation enabled by digital communication of content. Publishers, subscription agents, researchers and libraries are all looking into ways to create added value by linking related information entities, as such presenting the information within a broader context estimated to be relevant to the users of the information. This is the first of two articles in D-Lib Magazine on this topic. This first part describes the current state-of-the-art and contrasts various approaches to the problem. It identifies static and dynamic linking solutions as well as open and closed linking frameworks. It also includes an extensive bibliography. The second part, SFX, a Generic Linking Solution describes a system that we have developed for linking in a hybrid working environment. The creation of services linking related information entities is an area that is attracting an ever increasing interest in the ongoing development of the World Wide Web in general, and of research-related information systems in particular. Although most writings on electronic scientific communication have touted other benefits, such as the increase in communication speed, the possibility to exchange multimedia content and the absence of limitations on the length of research papers, currently both practice and theory point at linking services as being a major opportunity for improved communication of content. Publishers, subscription agents, researchers and libraries are all looking into ways to create added-value by linking related information entities, as such presenting the information within a broader context estimated to be relevant to the users of the information.
  16. Doerr, M.: ¬The CIDOC CRM, an ontological approach to schema heterogeneity (2005) 0.01
    0.014998848 = product of:
      0.06999462 = sum of:
        0.039667886 = weight(_text_:wide in 1662) [ClassicSimilarity], result of:
          0.039667886 = score(doc=1662,freq=2.0), product of:
            0.13505316 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.030480823 = queryNorm
            0.29372054 = fieldWeight in 1662, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.046875 = fieldNorm(doc=1662)
        0.021520529 = weight(_text_:web in 1662) [ClassicSimilarity], result of:
          0.021520529 = score(doc=1662,freq=2.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.21634221 = fieldWeight in 1662, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.046875 = fieldNorm(doc=1662)
        0.0088062035 = weight(_text_:information in 1662) [ClassicSimilarity], result of:
          0.0088062035 = score(doc=1662,freq=4.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.16457605 = fieldWeight in 1662, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.046875 = fieldNorm(doc=1662)
      0.21428572 = coord(3/14)
    
    Abstract
    The creation of the World Wide Web has had a profound impact an the ease with which information can be distributed and presented. Now with more and more information becoming available, there is an increasing demand for targeted global search, comparative studies, data transfer and data migration between heterogeneous sources of cultural and scholarly contents. This requires interoperability not only at the encoding level - a task solved well by XML for instance - but also at the more complex semantics level, where lie the characteristics of the domain. In the meanwhile, the reality of semantic interoperability is getting frustrating. In the cultural area alone, dozens of "standard" and hundreds of proprietary metadata and data structures exist, as well as hundreds of terminology systems. Core systems like the Dublin Core represent a common denominator by far too small to fulfil advanced requirements. Overstretching its already limited semantics in order to capture complex contents leads to further loss of meaning.
  17. Baker, T.: ¬A grammar of Dublin Core (2000) 0.01
    0.01490554 = product of:
      0.05216939 = sum of:
        0.026445258 = weight(_text_:wide in 1236) [ClassicSimilarity], result of:
          0.026445258 = score(doc=1236,freq=2.0), product of:
            0.13505316 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.030480823 = queryNorm
            0.1958137 = fieldWeight in 1236, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.03125 = fieldNorm(doc=1236)
        0.014347021 = weight(_text_:web in 1236) [ClassicSimilarity], result of:
          0.014347021 = score(doc=1236,freq=2.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.14422815 = fieldWeight in 1236, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.03125 = fieldNorm(doc=1236)
        0.0058708023 = weight(_text_:information in 1236) [ClassicSimilarity], result of:
          0.0058708023 = score(doc=1236,freq=4.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.10971737 = fieldWeight in 1236, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.03125 = fieldNorm(doc=1236)
        0.005506308 = product of:
          0.016518923 = sum of:
            0.016518923 = weight(_text_:22 in 1236) [ClassicSimilarity], result of:
              0.016518923 = score(doc=1236,freq=2.0), product of:
                0.10673865 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030480823 = queryNorm
                0.15476047 = fieldWeight in 1236, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1236)
          0.33333334 = coord(1/3)
      0.2857143 = coord(4/14)
    
    Abstract
    Dublin Core is often presented as a modern form of catalog card -- a set of elements (and now qualifiers) that describe resources in a complete package. Sometimes it is proposed as an exchange format for sharing records among multiple collections. The founding principle that "every element is optional and repeatable" reinforces the notion that a Dublin Core description is to be taken as a whole. This paper, in contrast, is based on a much different premise: Dublin Core is a language. More precisely, it is a small language for making a particular class of statements about resources. Like natural languages, it has a vocabulary of word-like terms, the two classes of which -- elements and qualifiers -- function within statements like nouns and adjectives; and it has a syntax for arranging elements and qualifiers into statements according to a simple pattern. Whenever tourists order a meal or ask directions in an unfamiliar language, considerate native speakers will spontaneously limit themselves to basic words and simple sentence patterns along the lines of "I am so-and-so" or "This is such-and-such". Linguists call this pidginization. In such situations, a small phrase book or translated menu can be most helpful. By analogy, today's Web has been called an Internet Commons where users and information providers from a wide range of scientific, commercial, and social domains present their information in a variety of incompatible data models and description languages. In this context, Dublin Core presents itself as a metadata pidgin for digital tourists who must find their way in this linguistically diverse landscape. Its vocabulary is small enough to learn quickly, and its basic pattern is easily grasped. It is well-suited to serve as an auxiliary language for digital libraries. This grammar starts by defining terms. It then follows a 200-year-old tradition of English grammar teaching by focusing on the structure of single statements. It concludes by looking at the growing dictionary of Dublin Core vocabulary terms -- its registry, and at how statements can be used to build the metadata equivalent of paragraphs and compositions -- the application profile.
    Date
    26.12.2011 14:01:22
  18. Mirizzi, R.: Exploratory browsing in the Web of Data (2011) 0.01
    0.0148904305 = product of:
      0.069488674 = sum of:
        0.046971425 = weight(_text_:web in 4803) [ClassicSimilarity], result of:
          0.046971425 = score(doc=4803,freq=28.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.47219574 = fieldWeight in 4803, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4803)
        0.0072647473 = weight(_text_:information in 4803) [ClassicSimilarity], result of:
          0.0072647473 = score(doc=4803,freq=8.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.13576832 = fieldWeight in 4803, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4803)
        0.015252502 = weight(_text_:retrieval in 4803) [ClassicSimilarity], result of:
          0.015252502 = score(doc=4803,freq=4.0), product of:
            0.092201896 = queryWeight, product of:
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.030480823 = queryNorm
            0.16542503 = fieldWeight in 4803, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              3.024915 = idf(docFreq=5836, maxDocs=44218)
              0.02734375 = fieldNorm(doc=4803)
      0.21428572 = coord(3/14)
    
    Abstract
    Thanks to the recent Linked Data initiative, the foundations of the Semantic Web have been built. Shared, open and linked RDF datasets give us the possibility to exploit both the strong theoretical results and the robust technologies and tools developed since the seminal paper in the Semantic Web appeared in 2001. In a simplistic way, we may think at the Semantic Web as a ultra large distributed database we can query to get information coming from different sources. In fact, every dataset exposes a SPARQL endpoint to make the data accessible through exact queries. If we know the URI of the famous actress Nicole Kidman in DBpedia we may retrieve all the movies she acted with a simple SPARQL query. Eventually we may aggregate this information with users ratings and genres from IMDB. Even though these are very exciting results and applications, there is much more behind the curtains. Datasets come with the description of their schema structured in an ontological way. Resources refer to classes which are in turn organized in well structured and rich ontologies. Exploiting also this further feature we go beyond the notion of a distributed database and we can refer to the Semantic Web as a distributed knowledge base. If in our knowledge base we have that Paris is located in France (ontological level) and that Moulin Rouge! is set in Paris (data level) we may query the Semantic Web (interpreted as a set of interconnected datasets and related ontologies) to return all the movies starred by Nicole Kidman set in France and Moulin Rouge! will be in the final result set. The ontological level makes possible to infer new relations among data.
    The Linked Data initiative and the state of the art in semantic technologies led off all brand new search and mash-up applications. The basic idea is to have smarter lookup services for a huge, distributed and social knowledge base. All these applications catch and (re)propose, under a semantic data perspective, the view of the classical Web as a distributed collection of documents to retrieve. The interlinked nature of the Web, and consequently of the Semantic Web, is exploited (just) to collect and aggregate data coming from different sources. Of course, this is a big step forward in search and Web technologies, but if we limit our investi- gation to retrieval tasks, we miss another important feature of the current Web: browsing and in particular exploratory browsing (a.k.a. exploratory search). Thanks to its hyperlinked nature, the Web defined a new way of browsing documents and knowledge: selection by lookup, navigation and trial-and-error tactics were, and still are, exploited by users to search for relevant information satisfying some initial requirements. The basic assumptions behind a lookup search, typical of Information Retrieval (IR) systems, are no more valid in an exploratory browsing context. An IR system, such as a search engine, assumes that: the user has a clear picture of what she is looking for ; she knows the terminology of the specific knowledge space. On the other side, as argued in, the main challenges in exploratory search can be summarized as: support querying and rapid query refinement; other facets and metadata-based result filtering; leverage search context; support learning and understanding; other visualization to support insight/decision making; facilitate collaboration. In Section 3 we will show two applications for exploratory search in the Semantic Web addressing some of the above challenges.
    Theme
    Semantic Web
  19. Cregan, A.: ¬An OWL DL construction for the ISO Topic Map Data Model (2005) 0.01
    0.01485169 = product of:
      0.069307886 = sum of:
        0.033056572 = weight(_text_:wide in 4718) [ClassicSimilarity], result of:
          0.033056572 = score(doc=4718,freq=2.0), product of:
            0.13505316 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.030480823 = queryNorm
            0.24476713 = fieldWeight in 4718, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4718)
        0.03106221 = weight(_text_:web in 4718) [ClassicSimilarity], result of:
          0.03106221 = score(doc=4718,freq=6.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.3122631 = fieldWeight in 4718, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4718)
        0.005189105 = weight(_text_:information in 4718) [ClassicSimilarity], result of:
          0.005189105 = score(doc=4718,freq=2.0), product of:
            0.053508412 = queryWeight, product of:
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.030480823 = queryNorm
            0.09697737 = fieldWeight in 4718, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              1.7554779 = idf(docFreq=20772, maxDocs=44218)
              0.0390625 = fieldNorm(doc=4718)
      0.21428572 = coord(3/14)
    
    Abstract
    Both Topic Maps and the W3C Semantic Web technologies are meta-level semantic maps describing relationships between information resources. Previous attempts at interoperability between XTM Topic Maps and RDF have proved problematic. The ISO's drafting of an explicit Topic Map Data Model [TMDM 05] combined with the advent of the W3C's XML and RDFbased Description Logic-equivalent Web Ontology Language [OWLDL 04] now provides the means for the construction of an unambiguous semantic model to represent Topic Maps, in a form that is equivalent to a Description Logic representation. This paper describes the construction of the proposed TMDM ISO Topic Map Standard in OWL DL (Description Logic equivalent) form. The construction is claimed to exactly match the features of the proposed TMDM. The intention is that the topic map constructs described herein, once officially published on the world-wide web, may be used by Topic Map authors to construct their Topic Maps in OWL DL. The advantage of OWL DL Topic Map construction over XTM, the existing XML-based DTD standard, is that OWL DL allows many constraints to be explicitly stated. OWL DL's suite of tools, although currently still somewhat immature, will provide the means for both querying and enforcing constraints. This goes a long way towards fulfilling the requirements for a Topic Map Query Language (TMQL) and Constraint Language (TMCL), which the Topic Map Community may choose to expend effort on extending. Additionally, OWL DL has a clearly defined formal semantics (Description Logic ref)
  20. Bechhofer, S.; Harmelen, F. van; Hendler, J.; Horrocks, I.; McGuinness, D.L.; Patel-Schneider, P.F.; Stein, L.A.: OWL Web Ontology Language Reference (2004) 0.01
    0.014631543 = product of:
      0.1024208 = sum of:
        0.046279203 = weight(_text_:wide in 4684) [ClassicSimilarity], result of:
          0.046279203 = score(doc=4684,freq=2.0), product of:
            0.13505316 = queryWeight, product of:
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.030480823 = queryNorm
            0.342674 = fieldWeight in 4684, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.4307585 = idf(docFreq=1430, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4684)
        0.056141596 = weight(_text_:web in 4684) [ClassicSimilarity], result of:
          0.056141596 = score(doc=4684,freq=10.0), product of:
            0.09947448 = queryWeight, product of:
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.030480823 = queryNorm
            0.5643819 = fieldWeight in 4684, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              3.2635105 = idf(docFreq=4597, maxDocs=44218)
              0.0546875 = fieldNorm(doc=4684)
      0.14285715 = coord(2/14)
    
    Abstract
    The Web Ontology Language OWL is a semantic markup language for publishing and sharing ontologies on the World Wide Web. OWL is developed as a vocabulary extension of RDF (the Resource Description Framework) and is derived from the DAML+OIL Web Ontology Language. This document contains a structured informal description of the full set of OWL language constructs and is meant to serve as a reference for OWL users who want to construct OWL ontologies.
    Theme
    Semantic Web

Years

Languages

Types

  • a 494
  • i 39
  • r 27
  • m 21
  • x 21
  • s 19
  • n 17
  • b 9
  • p 6
  • l 1
  • More… Less…

Themes

Subjects