Search (43 results, page 1 of 3)

  • × type_ss:"a"
  • × type_ss:"el"
  • × year_i:[2000 TO 2010}
  1. Heflin, J.; Hendler, J.: Semantic interoperability on the Web (2000) 0.01
    0.0074971514 = product of:
      0.029988606 = sum of:
        0.02024465 = product of:
          0.06073395 = sum of:
            0.06073395 = weight(_text_:problem in 759) [ClassicSimilarity], result of:
              0.06073395 = score(doc=759,freq=4.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.46424055 = fieldWeight in 759, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=759)
          0.33333334 = coord(1/3)
        0.009743956 = product of:
          0.029231867 = sum of:
            0.029231867 = weight(_text_:22 in 759) [ClassicSimilarity], result of:
              0.029231867 = score(doc=759,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.2708308 = fieldWeight in 759, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=759)
          0.33333334 = coord(1/3)
      0.25 = coord(2/8)
    
    Abstract
    XML will have a profound impact on the way data is exchanged on the Internet. An important feature of this language is the separation of content from presentation, which makes it easier to select and/or reformat the data. However, due to the likelihood of numerous industry and domain specific DTDs, those who wish to integrate information will still be faced with the problem of semantic interoperability. In this paper we discuss why this problem is not solved by XML, and then discuss why the Resource Description Framework is only a partial solution. We then present the SHOE language, which we feel has many of the features necessary to enable a semantic web, and describe an existing set of tools that make it easy to use the language.
    Date
    11. 5.2013 19:22:18
  2. Foerster, H. von; Müller, A.; Müller, K.H.: Rück- und Vorschauen : Heinz von Foerster im Gespräch mit Albert Müller und Karl H. Müller (2001) 0.00
    0.004111523 = product of:
      0.016446091 = sum of:
        0.012270111 = product of:
          0.03681033 = sum of:
            0.03681033 = weight(_text_:problem in 5988) [ClassicSimilarity], result of:
              0.03681033 = score(doc=5988,freq=8.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.28137225 = fieldWeight in 5988, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=5988)
          0.33333334 = coord(1/3)
        0.004175981 = product of:
          0.012527943 = sum of:
            0.012527943 = weight(_text_:22 in 5988) [ClassicSimilarity], result of:
              0.012527943 = score(doc=5988,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.116070345 = fieldWeight in 5988, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0234375 = fieldNorm(doc=5988)
          0.33333334 = coord(1/3)
      0.25 = coord(2/8)
    
    Content
    Darin: ... "Eine Idee, mit der wir uns beschäftigt haben, kommt jetzt langsam wirklich heraus. Wie rechnet man innerhalb einer semantischen Struktur? Wir haben das so gesehen, daß jedes Wort, jeder Begriff so ausschaut wie ein vielfältiges Element, das nach allen Richtungen seine Konnektivitäten ausstreckt und mit anderen solchen vielfältigen Elementen in Verbindung bringt. Und die Operationen bestehen darin, neue Verbindungen zu finden, die grammatisch kontrolliert werden und als Sprache herauskommen, aber konzeptuell konnektiert, so daß sie verbunden sind durch eine semantische interne Struktur. Das heißt, jeder Begriff ist für uns ein vielfältiger Rechner, der sich mit anderen Rechnern in Verbindung setzt. Damals hat das niemand verstanden, vielleicht habe ich es auch nicht gut dargestellt. Aber heute taucht das überall auf, semantic computation, mit lauter parallelen Maschinen, die alle gleichzeitig arbeiten und ihre Verbindungen herstellen. Unser Problem war damals schon: könnte man irgendetwas machen, um in natürlichen Sprachen mit einer Maschine sprechen zu können.
    Noch einige Schritte weiter zurück. Oft haben mich Bibliothekare angesprochen, wie sollte man eine Bibliothek aufbauen? Wir schauen, sagten sie, in eine Bibliothek so hinein, als wäre sie wie ein Gedächtnis. "Das ist schön, aber wissen Sie, wie das Gedächtnis funktioniert? "Nein, aber viele Leute sagen, das Gedächtnis arbeitet wie eine große Bibliothek. Man muß nur hineingreifen und das richtige Buch finden. "Das ist alles wunderschön und sehr lieb, aber wissen Sie, die Leute, die ein Buch suchen, suchen es ja nur, weil sie ein Problem haben und hoffen, in dem Buch die Antwort für das Problem zu finden. Das Buch ist nur ein Zwischenträger von einer Frage und einer vielleicht in dem Buch zu findenden Antwort. Aber das Buch ist nicht die Antwort. "Aha, wie stellen Sie sich das vor? Wir sollten das Problem so sehen, daß die Inhalte der Bücher, die semantische Struktur - wenn man jetzt diesen Ausdruck wieder verwenden möchte - dieser Bücher in einem System sitzt, sodaß ich in diese semantische Struktur mit meiner Frage einsteigen kann, und mir die semantische Struktur dieses Systems sagt, dann mußt du Karl Müllers Arbeiten über Symbole lesen, dann wirst du wissen, was du suchst. Ich wüßte aber von vornherein überhaupt nicht, wer der Karl Müller ist, daß er über Symbole geschrieben hat, etc., aber das System kann mir das liefern.
    Date
    10. 9.2006 17:22:54
  3. Bradford, R.B.: Relationship discovery in large text collections using Latent Semantic Indexing (2006) 0.00
    0.003437012 = product of:
      0.013748048 = sum of:
        0.008180073 = product of:
          0.02454022 = sum of:
            0.02454022 = weight(_text_:problem in 1163) [ClassicSimilarity], result of:
              0.02454022 = score(doc=1163,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.1875815 = fieldWeight in 1163, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1163)
          0.33333334 = coord(1/3)
        0.005567975 = product of:
          0.016703924 = sum of:
            0.016703924 = weight(_text_:22 in 1163) [ClassicSimilarity], result of:
              0.016703924 = score(doc=1163,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.15476047 = fieldWeight in 1163, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1163)
          0.33333334 = coord(1/3)
      0.25 = coord(2/8)
    
    Abstract
    This paper addresses the problem of information discovery in large collections of text. For users, one of the key problems in working with such collections is determining where to focus their attention. In selecting documents for examination, users must be able to formulate reasonably precise queries. Queries that are too broad will greatly reduce the efficiency of information discovery efforts by overwhelming the users with peripheral information. In order to formulate efficient queries, a mechanism is needed to automatically alert users regarding potentially interesting information contained within the collection. This paper presents the results of an experiment designed to test one approach to generation of such alerts. The technique of latent semantic indexing (LSI) is used to identify relationships among entities of interest. Entity extraction software is used to pre-process the text of the collection so that the LSI space contains representation vectors for named entities in addition to those for individual terms. In the LSI space, the cosine of the angle between the representation vectors for two entities captures important information regarding the degree of association of those two entities. For appropriate choices of entities, determining the entity pairs with the highest mutual cosine values yields valuable information regarding the contents of the text collection. The test database used for the experiment consists of 150,000 news articles. The proposed approach for alert generation is tested using a counterterrorism analysis example. The approach is shown to have significant potential for aiding users in rapidly focusing on information of potential importance in large text collections. The approach also has value in identifying possible use of aliases.
    Source
    Proceedings of the Fourth Workshop on Link Analysis, Counterterrorism, and Security, SIAM Data Mining Conference, Bethesda, MD, 20-22 April, 2006. [http://www.siam.org/meetings/sdm06/workproceed/Link%20Analysis/15.pdf]
  4. Wake, S.; Nicholson, D.: HILT: High-Level Thesaurus Project : building consensus for interoperable subject access across communities (2001) 0.00
    0.0030675277 = product of:
      0.024540221 = sum of:
        0.024540221 = product of:
          0.07362066 = sum of:
            0.07362066 = weight(_text_:problem in 1224) [ClassicSimilarity], result of:
              0.07362066 = score(doc=1224,freq=18.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.5627445 = fieldWeight in 1224, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.03125 = fieldNorm(doc=1224)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    This article provides an overview of the work carried out by the HILT Project <http://hilt.cdlr.strath.ac.uk> in making recommendations towards interoperable subject access, or cross-searching and browsing distributed services amongst the archives, libraries, museums and electronic services sectors. The article details consensus achieved at the 19 June 2001 HILT Workshop and discusses the HILT Stakeholder Survey. In 1999 Péter Jascó wrote that "savvy searchers" are asking for direction. Three years later the scenario he describes, that of searchers cross-searching databases where the subject vocabulary used in each case is different, still rings true. Jascó states that, in many cases, databases do not offer the necessary aids required to use the "preferred terms of the subject-controlled vocabulary". The databases to which Jascó refers are Dialog and DataStar. However, the situation he describes applies as well to the area that HILT is researching: that of cross-searching and browsing by subject across databases and catalogues in archives, libraries, museums and online information services. So how does a user access information on a particular subject when it is indexed across a multitude of services under different, but quite often similar, subject terms? Also, if experienced searchers are having problems, what about novice searchers? As information professionals, it is our role to investigate such problems and recommend solutions. Although there is no hard empirical evidence one way or another, HILT participants agree that the problem for users attempting to search across databases is real. There is a strong likelihood that users are disadvantaged by the use of different subject terminology combined with a multitude of different practices taking place within the archive, library, museums and online communities. Arguably, failure to address this problem of interoperability undermines the value of cross-searching and browsing facilities, and wastes public money because relevant resources are 'hidden' from searchers. HILT is charged with analysing this broad problem through qualitative methods, with the main aim of presenting a set of recommendations on how to make it easier to cross-search and browse distributed services. Because this is a very large problem composed of many strands, HILT recognizes that any proposed solutions must address a host of issues. Recommended solutions must be affordable, sustainable, politically acceptable, useful, future-proof and international in scope. It also became clear to the HILT team that progress toward finding solutions to the interoperability problem could only be achieved through direct dialogue with other parties keen to solve this problem, and that the problem was as much about consensus building as it was about finding a solution. This article describes how HILT approached the cross-searching problem; how it investigated the nature of the problem, detailing results from the HILT Stakeholder Survey; and how it achieved consensus through the recent HILT Workshop.
  5. Van der Veer Martens, B.: Do citation systems represent theories of truth? (2001) 0.00
    0.0024607205 = product of:
      0.019685764 = sum of:
        0.019685764 = product of:
          0.059057288 = sum of:
            0.059057288 = weight(_text_:22 in 3925) [ClassicSimilarity], result of:
              0.059057288 = score(doc=3925,freq=4.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.54716086 = fieldWeight in 3925, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3925)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    22. 7.2006 15:22:28
  6. Rudner, L.: Who is going to mine digital library resources? : anf how? (2000) 0.00
    0.0021069439 = product of:
      0.01685555 = sum of:
        0.01685555 = product of:
          0.05056665 = sum of:
            0.05056665 = weight(_text_:29 in 6800) [ClassicSimilarity], result of:
              0.05056665 = score(doc=6800,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.46638384 = fieldWeight in 6800, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.09375 = fieldNorm(doc=6800)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    26.12.2011 16:38:29
  7. Qin, J.; Paling, S.: Converting a controlled vocabulary into an ontology : the case of GEM (2001) 0.00
    0.0020879905 = product of:
      0.016703924 = sum of:
        0.016703924 = product of:
          0.05011177 = sum of:
            0.05011177 = weight(_text_:22 in 3895) [ClassicSimilarity], result of:
              0.05011177 = score(doc=3895,freq=2.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.46428138 = fieldWeight in 3895, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.09375 = fieldNorm(doc=3895)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    24. 8.2005 19:20:22
  8. Aitken, S.; Reid, S.: Evaluation of an ontology-based information retrieval tool (2000) 0.00
    0.0020450184 = product of:
      0.016360147 = sum of:
        0.016360147 = product of:
          0.04908044 = sum of:
            0.04908044 = weight(_text_:problem in 2862) [ClassicSimilarity], result of:
              0.04908044 = score(doc=2862,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.375163 = fieldWeight in 2862, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0625 = fieldNorm(doc=2862)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Content
    Beitrag für: Workshop on the Applications of Ontologies and Problem-Solving Methods, (eds) Gómez-Pérez, A., Benjamins, V.R., Guarino, N., and Uschold, M. European Conference on Artificial Intelligence 2000, Berlin.
  9. Griffiths, T.L.; Steyvers, M.: ¬A probabilistic approach to semantic representation (2002) 0.00
    0.0019864459 = product of:
      0.015891567 = sum of:
        0.015891567 = product of:
          0.047674697 = sum of:
            0.047674697 = weight(_text_:29 in 3671) [ClassicSimilarity], result of:
              0.047674697 = score(doc=3671,freq=4.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.43971092 = fieldWeight in 3671, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0625 = fieldNorm(doc=3671)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    29. 6.2015 14:55:01
    29. 6.2015 16:09:05
  10. Doerr, M.: Semantic problems of thesaurus mapping (2001) 0.00
    0.0018075579 = product of:
      0.014460463 = sum of:
        0.014460463 = product of:
          0.04338139 = sum of:
            0.04338139 = weight(_text_:problem in 5902) [ClassicSimilarity], result of:
              0.04338139 = score(doc=5902,freq=4.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.33160037 = fieldWeight in 5902, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5902)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    With networked information access to heterogeneous data sources, the problem of terminology provision and interoperability of controlled vocabulary schemes such as thesauri becomes increasingly urgent. Solutions are needed to improve the performance of full-text retrieval systems and to guide the design of controlled terminology schemes for use in structured data, including metadata. Thesauri are created in different languages, with different scope and points of view and at different levels of abstraction and detail, to accomodate access to a specific group of collections. In any wider search accessing distributed collections, the user would like to start with familiar terminology and let the system find out the correspondences to other terminologies in order to retrieve equivalent results from all addressed collections. This paper investigates possible semantic differences that may hinder the unambiguous mapping and transition from one thesaurus to another. It focusses on the differences of meaning of terms and their relations as intended by their creators for indexing and querying a specific collection, in contrast to methods investigating the statistical relevance of terms for objects in a collection. It develops a notion of optimal mapping, paying particular attention to the intellectual quality of mappings between terms from different vocabularies and to problems of polysemy. Proposals are made to limit the vagueness introduced by the transition from one vocabulary to another. The paper shows ways in which thesaurus creators can improve their methodology to meet the challenges of networked access of distributed collections created under varying conditions. For system implementers, the discussion will lead to a better understanding of the complexity of the problem
  11. Janée, G.; Frew, J.; Hill, L.L.: Issues in georeferenced digital libraries (2004) 0.00
    0.001789391 = product of:
      0.014315128 = sum of:
        0.014315128 = product of:
          0.042945385 = sum of:
            0.042945385 = weight(_text_:problem in 1165) [ClassicSimilarity], result of:
              0.042945385 = score(doc=1165,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.3282676 = fieldWeight in 1165, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1165)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    Based on a decade's experience with the Alexandria Digital Library Project, seven issues are presented that arise in creating georeferenced digital libraries, and that appear to be intrinsic to the problem of creating any library-like information system that operates on georeferenced and geospatial resources. The first and foremost issue is providing discovery of georeferenced resources. Related to discovery are the issues of gazetteer integration and specialized ranking of search results. Strong data typing and scalability are implementation issues. Providing spatial context is a critical user interface issue. Finally, sophisticated resource access mechanisms are necessary to operate on geospatial resources.
  12. Shechtman, N.; Chung, M.; Roschelle, J.: Supporting member collaboration in the Math Tools digital library : a formative user study (2004) 0.00
    0.0015337638 = product of:
      0.012270111 = sum of:
        0.012270111 = product of:
          0.03681033 = sum of:
            0.03681033 = weight(_text_:problem in 1163) [ClassicSimilarity], result of:
              0.03681033 = score(doc=1163,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.28137225 = fieldWeight in 1163, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1163)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    In this paper, we discuss a user study done at the formative stage of development of a Math Tools developers' community. The Math Tools digital library, which aims to collect software tools to support K-12 and university mathematics instruction, has two synergistic purposes. One is to support federated search and the other is to create a community of practice in which developers and users can work together. While much research has explored the technical problem of federated search, there has been little investigation into how to grow a creative, working community around a digital library. To this end, we surveyed and interviewed members of the Math Tools community in order to elicit concerns and priorities. These data led to rich descriptions of the teachers, developers, and researchers who comprise this community. Insights from these descriptions were then used to inform the creation of a set of metaphors and design principles that the Math Tools team could use in their continuing design work.
  13. Combs, A.; Krippner, S.: Collective consciousness and the social brain (2008) 0.00
    0.0015337638 = product of:
      0.012270111 = sum of:
        0.012270111 = product of:
          0.03681033 = sum of:
            0.03681033 = weight(_text_:problem in 5622) [ClassicSimilarity], result of:
              0.03681033 = score(doc=5622,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.28137225 = fieldWeight in 5622, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5622)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    This paper discusses supportive neurological and social evidence for 'collective consciousness', here understood as a shared sense of being together with others in a single or unified experience. Mirror neurons in the premotor and posterior parietal cortices respond to the intentions as well as the actions of other individuals. There are also mirror neurons in the anterior insula and anterior cingulate cortices which have been implicated in empathy. Many authors have considered the likely role of such mirror systems in the development of uniquely human aspects of sociality including language. Though not without criticism, Menant has made the case that mirror-neuron assisted exchanges aided the original advent of self-consciousness and intersubjectivity. Combining these ideas with social mirror theory it is not difficult to imagine the creation of similar dynamical patterns in the emotional and even cognitive neuronal activity of individuals in human groups, creating a feeling in which the participating members experience a unified sense of consciousness. Such instances pose a kind of 'binding problem' in which participating individuals exhibit a degree of 'entanglement'.
  14. Hammond, T.; Hannay, T.; Lund, B.; Flack, M.: Social bookmarking tools (II) : a case study - Connotea (2005) 0.00
    0.0012781365 = product of:
      0.010225092 = sum of:
        0.010225092 = product of:
          0.030675275 = sum of:
            0.030675275 = weight(_text_:problem in 1189) [ClassicSimilarity], result of:
              0.030675275 = score(doc=1189,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.23447686 = fieldWeight in 1189, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1189)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    Connotea is a free online reference management and social bookmarking service for scientists created by Nature Publishing Group. While somewhat experimental in nature, Connotea already has a large and growing number of users, and is a real, fully functioning service. The label 'experimental' is not meant to imply that the service is any way ephemeral or esoteric, rather that the concept of social bookmarking itself and the application of that concept to reference management are both recent developments. Connotea is under active development, and we are still in the process of discovering how people will use it. In addition to Connotea being a free and public service, the core code is freely available under an open source license. Connotea was conceived from the outset as an online, social tool. Seeing the possibilities that del.icio.us was opening up for its users in the area of general web linking, we realised that scholarly reference management was a similar problem space. Connotea was designed and developed late in 2004, and soft-launched at the end of December 2004. Usage has grown over the past several months, to the point where there is now enough data in the system for interesting second-order effects to emerge. This paper will start by giving an overview of Connotea, and will outline the key concepts and describe its main features. We will then take the reader on a brief guided tour, show some of the aforementioned second-order effects, and end with a discussion of Connotea's likely future direction.
  15. Tudhope, D.; Alani, H.; Jones, C.: Augmenting thesaurus relationships : possibilities for retrieval (2001) 0.00
    0.0012781365 = product of:
      0.010225092 = sum of:
        0.010225092 = product of:
          0.030675275 = sum of:
            0.030675275 = weight(_text_:problem in 1520) [ClassicSimilarity], result of:
              0.030675275 = score(doc=1520,freq=2.0), product of:
                0.13082431 = queryWeight, product of:
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.030822188 = queryNorm
                0.23447686 = fieldWeight in 1520, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.244485 = idf(docFreq=1723, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1520)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    This paper discusses issues concerning the augmentation of thesaurus relationships, in light of new application possibilities for retrieval. We first discuss a case study that explored the retrieval potential of an augmented set of thesaurus relationships by specialising standard relationships into richer subtypes, in particular hierarchical geographical containment and the associative relationship. We then locate this work in a broader context by reviewing various attempts to build taxonomies of thesaurus relationships, and conclude by discussing the feasibility of hierarchically augmenting the core set of thesaurus relationships, particularly the associative relationship. We discuss the possibility of enriching the specification and semantics of Related Term (RT relationships), while maintaining compatibility with traditional thesauri via a limited hierarchical extension of the associative (and hierarchical) relationships. This would be facilitated by distinguishing the type of term from the (sub)type of relationship and explicitly specifying semantic categories for terms following a faceted approach. We first illustrate how hierarchical spatial relationships can be used to provide more flexible retrieval for queries incorporating place names in applications employing online gazetteers and geographical thesauri. We then employ a set of experimental scenarios to investigate key issues affecting use of the associative (RT) thesaurus relationships in semantic distance measures. Previous work has noted the potential of RTs in thesaurus search aids but also the problem of uncontrolled expansion of query term sets. Results presented in this paper suggest the potential for taking account of the hierarchical context of an RT link and specialisations of the RT relationship
  16. Decimal Classification Editorial Policy Committee (2002) 0.00
    0.0012303602 = product of:
      0.009842882 = sum of:
        0.009842882 = product of:
          0.029528644 = sum of:
            0.029528644 = weight(_text_:22 in 236) [ClassicSimilarity], result of:
              0.029528644 = score(doc=236,freq=4.0), product of:
                0.10793405 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.030822188 = queryNorm
                0.27358043 = fieldWeight in 236, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=236)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Abstract
    The Decimal Classification Editorial Policy Committee (EPC) held its Meeting 117 at the Library Dec. 3-5, 2001, with chair Andrea Stamm (Northwestern University) presiding. Through its actions at this meeting, significant progress was made toward publication of DDC unabridged Edition 22 in mid-2003 and Abridged Edition 14 in early 2004. For Edition 22, the committee approved the revisions to two major segments of the classification: Table 2 through 55 Iran (the first half of the geographic area table) and 900 History and geography. EPC approved updates to several parts of the classification it had already considered: 004-006 Data processing, Computer science; 340 Law; 370 Education; 510 Mathematics; 610 Medicine; Table 3 issues concerning treatment of scientific and technical themes, with folklore, arts, and printing ramifications at 398.2 - 398.3, 704.94, and 758; Table 5 and Table 6 Ethnic Groups and Languages (portions concerning American native peoples and languages); and tourism issues at 647.9 and 790. Reports on the results of testing the approved 200 Religion and 305-306 Social groups schedules were received, as was a progress report on revision work for the manual being done by Ross Trotter (British Library, retired). Revisions for Abridged Edition 14 that received committee approval included 010 Bibliography; 070 Journalism; 150 Psychology; 370 Education; 380 Commerce, communications, and transportation; 621 Applied physics; 624 Civil engineering; and 629.8 Automatic control engineering. At the meeting the committee received print versions of _DC&_ numbers 4 and 5. Primarily for the use of Dewey translators, these cumulations list changes, substantive and cosmetic, to DDC Edition 21 and Abridged Edition 13 for the period October 1999 - December 2001. EPC will hold its Meeting 118 at the Library May 15-17, 2002.
  17. Francu, V.: Does convenience trump accuracy? : the avatars of the UDC in Romania (2007) 0.00
    0.0012290506 = product of:
      0.009832405 = sum of:
        0.009832405 = product of:
          0.029497212 = sum of:
            0.029497212 = weight(_text_:29 in 544) [ClassicSimilarity], result of:
              0.029497212 = score(doc=544,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.27205724 = fieldWeight in 544, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=544)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Source
    Extensions and corrections to the UDC. 29(2007), S.263-272
  18. Goodchild, M.F.: ¬The Alexandria Digital Library Project : review, assessment, and prospects (2004) 0.00
    0.0012290506 = product of:
      0.009832405 = sum of:
        0.009832405 = product of:
          0.029497212 = sum of:
            0.029497212 = weight(_text_:29 in 1153) [ClassicSimilarity], result of:
              0.029497212 = score(doc=1153,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.27205724 = fieldWeight in 1153, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1153)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    26.12.2011 16:29:21
  19. Godby, C.J.; Young, J.A.; Childress, E.: ¬A repository of metadata crosswalks (2004) 0.00
    0.0012290506 = product of:
      0.009832405 = sum of:
        0.009832405 = product of:
          0.029497212 = sum of:
            0.029497212 = weight(_text_:29 in 1155) [ClassicSimilarity], result of:
              0.029497212 = score(doc=1155,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.27205724 = fieldWeight in 1155, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1155)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    26.12.2011 16:29:02
  20. Assem, M. van; Menken, M.R.; Schreiber, G.; Wielemaker, J.; Wielinga, B.: ¬A method for converting thesauri to RDF/OWL (2004) 0.00
    0.0012290506 = product of:
      0.009832405 = sum of:
        0.009832405 = product of:
          0.029497212 = sum of:
            0.029497212 = weight(_text_:29 in 4644) [ClassicSimilarity], result of:
              0.029497212 = score(doc=4644,freq=2.0), product of:
                0.108422816 = queryWeight, product of:
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.030822188 = queryNorm
                0.27205724 = fieldWeight in 4644, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5176873 = idf(docFreq=3565, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=4644)
          0.33333334 = coord(1/3)
      0.125 = coord(1/8)
    
    Date
    29. 7.2011 14:44:56

Languages

  • e 41
  • d 2
  • More… Less…