Search (25 results, page 1 of 2)

  • × author_ss:"Greenberg, J."
  1. White, H.C.; Carrier, S.; Thompson, A.; Greenberg, J.; Scherle, R.: ¬The Dryad Data Repository : a Singapore framework metadata architecture in a DSpace environment (2008) 0.02
    0.018633895 = product of:
      0.055901684 = sum of:
        0.055901684 = sum of:
          0.014351131 = weight(_text_:of in 2592) [ClassicSimilarity], result of:
            0.014351131 = score(doc=2592,freq=6.0), product of:
              0.06850986 = queryWeight, product of:
                1.5637573 = idf(docFreq=25162, maxDocs=44218)
                0.043811057 = queryNorm
              0.20947541 = fieldWeight in 2592, product of:
                2.4494898 = tf(freq=6.0), with freq of:
                  6.0 = termFreq=6.0
                1.5637573 = idf(docFreq=25162, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2592)
          0.041550554 = weight(_text_:22 in 2592) [ClassicSimilarity], result of:
            0.041550554 = score(doc=2592,freq=2.0), product of:
              0.15341885 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043811057 = queryNorm
              0.2708308 = fieldWeight in 2592, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0546875 = fieldNorm(doc=2592)
      0.33333334 = coord(1/3)
    
    Abstract
    This report presents recent metadata developments for Dryad, a digital repository hosting datasets underlying publications in the field of evolutionary biology. We review our efforts to bring the Dryad application profile into conformance with the Singapore Framework and discuss practical issues underlying the application profile implementation in a DSpace environment. The report concludes by outlining the next steps planned as Dryad moves into the next phase of development.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
  2. White, H.; Willis, C.; Greenberg, J.: HIVEing : the effect of a semantic web technology on inter-indexer consistency (2014) 0.02
    0.015811294 = product of:
      0.047433883 = sum of:
        0.047433883 = sum of:
          0.017754918 = weight(_text_:of in 1781) [ClassicSimilarity], result of:
            0.017754918 = score(doc=1781,freq=18.0), product of:
              0.06850986 = queryWeight, product of:
                1.5637573 = idf(docFreq=25162, maxDocs=44218)
                0.043811057 = queryNorm
              0.25915858 = fieldWeight in 1781, product of:
                4.2426405 = tf(freq=18.0), with freq of:
                  18.0 = termFreq=18.0
                1.5637573 = idf(docFreq=25162, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1781)
          0.029678967 = weight(_text_:22 in 1781) [ClassicSimilarity], result of:
            0.029678967 = score(doc=1781,freq=2.0), product of:
              0.15341885 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043811057 = queryNorm
              0.19345059 = fieldWeight in 1781, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1781)
      0.33333334 = coord(1/3)
    
    Abstract
    Purpose - The purpose of this paper is to examine the effect of the Helping Interdisciplinary Vocabulary Engineering (HIVE) system on the inter-indexer consistency of information professionals when assigning keywords to a scientific abstract. This study examined first, the inter-indexer consistency of potential HIVE users; second, the impact HIVE had on consistency; and third, challenges associated with using HIVE. Design/methodology/approach - A within-subjects quasi-experimental research design was used for this study. Data were collected using a task-scenario based questionnaire. Analysis was performed on consistency results using Hooper's and Rolling's inter-indexer consistency measures. A series of t-tests was used to judge the significance between consistency measure results. Findings - Results suggest that HIVE improves inter-indexing consistency. Working with HIVE increased consistency rates by 22 percent (Rolling's) and 25 percent (Hooper's) when selecting relevant terms from all vocabularies. A statistically significant difference exists between the assignment of free-text keywords and machine-aided keywords. Issues with homographs, disambiguation, vocabulary choice, and document structure were all identified as potential challenges. Research limitations/implications - Research limitations for this study can be found in the small number of vocabularies used for the study. Future research will include implementing HIVE into the Dryad Repository and studying its application in a repository system. Originality/value - This paper showcases several features used in HIVE system. By using traditional consistency measures to evaluate a semantic web technology, this paper emphasizes the link between traditional indexing and next generation machine-aided indexing (MAI) tools.
    Source
    Journal of documentation. 70(2014) no.3, S.307-329
  3. Shoffner, M.; Greenberg, J.; Kramer-Duffield, J.; Woodbury, D.: Web 2.0 semantic systems : collaborative learning in science (2008) 0.01
    0.014026793 = product of:
      0.042080376 = sum of:
        0.042080376 = sum of:
          0.018337203 = weight(_text_:of in 2661) [ClassicSimilarity], result of:
            0.018337203 = score(doc=2661,freq=30.0), product of:
              0.06850986 = queryWeight, product of:
                1.5637573 = idf(docFreq=25162, maxDocs=44218)
                0.043811057 = queryNorm
              0.26765788 = fieldWeight in 2661, product of:
                5.477226 = tf(freq=30.0), with freq of:
                  30.0 = termFreq=30.0
                1.5637573 = idf(docFreq=25162, maxDocs=44218)
                0.03125 = fieldNorm(doc=2661)
          0.023743173 = weight(_text_:22 in 2661) [ClassicSimilarity], result of:
            0.023743173 = score(doc=2661,freq=2.0), product of:
              0.15341885 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043811057 = queryNorm
              0.15476047 = fieldWeight in 2661, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=2661)
      0.33333334 = coord(1/3)
    
    Abstract
    The basic goal of education within a discipline is to transform a novice into an expert. This entails moving the novice toward the "semantic space" that the expert inhabits-the space of concepts, meanings, vocabularies, and other intellectual constructs that comprise the discipline. Metadata is significant to this goal in digitally mediated education environments. Encoding the experts' semantic space not only enables the sharing of semantics among discipline scientists, but also creates an environment that bridges the semantic gap between the common vocabulary of the novice and the granular descriptive language of the seasoned scientist (Greenberg, et al, 2005). Developments underlying the Semantic Web, where vocabularies are formalized in the Web Ontology Language (OWL), and Web 2.0 approaches of user-generated folksonomies provide an infrastructure for linking vocabulary systems and promoting group learning via metadata literacy. Group learning is a pedagogical approach to teaching that harnesses the phenomenon of "collective intelligence" to increase learning by means of collaboration. Learning a new semantic system can be daunting for a novice, and yet it is integral to advance one's knowledge in a discipline and retain interest. These ideas are key to the "BOT 2.0: Botany through Web 2.0, the Memex and Social Learning" project (Bot 2.0).72 Bot 2.0 is a collaboration involving the North Carolina Botanical Garden, the UNC SILS Metadata Research center, and the Renaissance Computing Institute (RENCI). Bot 2.0 presents a curriculum utilizing a memex as a way for students to link and share digital information, working asynchronously in an environment beyond the traditional classroom. Our conception of a memex is not a centralized black box but rather a flexible, distributed framework that uses the most salient and easiest-to-use collaborative platforms (e.g., Facebook, Flickr, wiki and blog technology) for personal information management. By meeting students "where they live" digitally, we hope to attract students to the study of botanical science. A key aspect is to teach students scientific terminology and about the value of metadata, an inherent function in several of the technologies and in the instructional approach we are utilizing. This poster will report on a study examining the value of both folksonomies and taxonomies for post-secondary college students learning plant identification. Our data is drawn from a curriculum involving a virtual independent learning portion and a "BotCamp" weekend at UNC, where students work with digital plan specimens that they have captured. Results provide some insight into the importance of collaboration and shared vocabulary for gaining confidence and for student progression from novice to expert in botany.
    Source
    Metadata for semantic and social applications : proceedings of the International Conference on Dublin Core and Metadata Applications, Berlin, 22 - 26 September 2008, DC 2008: Berlin, Germany / ed. by Jane Greenberg and Wolfgang Klas
  4. Willis, C.; Greenberg, J.; White, H.: Analysis and synthesis of metadata goals for scientific data (2012) 0.01
    0.013819531 = product of:
      0.04145859 = sum of:
        0.04145859 = sum of:
          0.01771542 = weight(_text_:of in 367) [ClassicSimilarity], result of:
            0.01771542 = score(doc=367,freq=28.0), product of:
              0.06850986 = queryWeight, product of:
                1.5637573 = idf(docFreq=25162, maxDocs=44218)
                0.043811057 = queryNorm
              0.25858206 = fieldWeight in 367, product of:
                5.2915025 = tf(freq=28.0), with freq of:
                  28.0 = termFreq=28.0
                1.5637573 = idf(docFreq=25162, maxDocs=44218)
                0.03125 = fieldNorm(doc=367)
          0.023743173 = weight(_text_:22 in 367) [ClassicSimilarity], result of:
            0.023743173 = score(doc=367,freq=2.0), product of:
              0.15341885 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.043811057 = queryNorm
              0.15476047 = fieldWeight in 367, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.03125 = fieldNorm(doc=367)
      0.33333334 = coord(1/3)
    
    Abstract
    The proliferation of discipline-specific metadata schemes contributes to artificial barriers that can impede interdisciplinary and transdisciplinary research. The authors considered this problem by examining the domains, objectives, and architectures of nine metadata schemes used to document scientific data in the physical, life, and social sciences. They used a mixed-methods content analysis and Greenberg's () metadata objectives, principles, domains, and architectural layout (MODAL) framework, and derived 22 metadata-related goals from textual content describing each metadata scheme. Relationships are identified between the domains (e.g., scientific discipline and type of data) and the categories of scheme objectives. For each strong correlation (>0.6), a Fisher's exact test for nonparametric data was used to determine significance (p < .05). Significant relationships were found between the domains and objectives of the schemes. Schemes describing observational data are more likely to have "scheme harmonization" (compatibility and interoperability with related schemes) as an objective; schemes with the objective "abstraction" (a conceptual model exists separate from the technical implementation) also have the objective "sufficiency" (the scheme defines a minimal amount of information to meet the needs of the community); and schemes with the objective "data publication" do not have the objective "element refinement." The analysis indicates that many metadata-driven goals expressed by communities are independent of scientific discipline or the type of data, although they are constrained by historical community practices and workflows as well as the technological environment at the time of scheme creation. The analysis reveals 11 fundamental metadata goals for metadata documenting scientific data in support of sharing research data across disciplines and domains. The authors report these results and highlight the need for more metadata-related research, particularly in the context of recent funding agency policy changes.
    Source
    Journal of the American Society for Information Science and Technology. 63(2012) no.8, S.1505-1520
  5. Greenberg, J.: Metadata generation : processes, people and tools (2003) 0.00
    0.004175565 = product of:
      0.012526695 = sum of:
        0.012526695 = product of:
          0.02505339 = sum of:
            0.02505339 = weight(_text_:of in 1251) [ClassicSimilarity], result of:
              0.02505339 = score(doc=1251,freq=14.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.36569026 = fieldWeight in 1251, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.0625 = fieldNorm(doc=1251)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Metadata generation is the act of creating or producing metadata. Generating good quality metadata in an efficient manner is essential for organizing and making accessible the growing number of rich resources available an the Web. The success of digital libraries, the sustenance of interoperability - as promoted by the Open Archives Initiative - and the evolution of Semantic Web all rely an efficient metadata generation. This article sketches a metadata generation framework that involves processes, people and tools. It also presents selected research initiatives and highlights the goals of the Metadata Generation Research Project.
    Source
    Bulletin of the American Society for Information Science. 29(2003) no.2, S.16-19
  6. Greenberg, J.: ¬A quantitative categorical analysis of metadata elements in image-applicable metadata schemes (2001) 0.00
    0.0039058835 = product of:
      0.01171765 = sum of:
        0.01171765 = product of:
          0.0234353 = sum of:
            0.0234353 = weight(_text_:of in 6529) [ClassicSimilarity], result of:
              0.0234353 = score(doc=6529,freq=16.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.34207192 = fieldWeight in 6529, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=6529)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    This article reports on a quantitative categorical analysis of metadata elements in the Dublin Core, VRA Core, REACH, and EAD metadata schemas, all of which can be used for organizing and describing images. The study found that each of the examined metadata schemas contains elements that support the discovery, use, authentication, and administration of images, and that the number and proportion of elements supporting functions in these classes varies per schema. The study introduces a new schema comparison methodology and explores the development of a class-oriented functional metadata schema for controlling images across multiple domains
    Source
    Journal of the American Society for Information Science and technology. 52(2001) no.11, S.917-924
  7. Newby, G.B.; Greenberg, J.; Jones, P.: Open source software development and Lotka's law : bibliometric patterns in programming (2003) 0.00
    0.0035509837 = product of:
      0.010652951 = sum of:
        0.010652951 = product of:
          0.021305902 = sum of:
            0.021305902 = weight(_text_:of in 5140) [ClassicSimilarity], result of:
              0.021305902 = score(doc=5140,freq=18.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.3109903 = fieldWeight in 5140, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5140)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Newby, Greenberg, and Jones analyze programming productivity of open source software by counting registered developers contributions found in the Linux Software Map and in Scourceforge. Using seven years of data from a subset of the Linux directory tree LSM data provided 4503 files with 3341 unique author names. The distribution follows Lotka's Law with an exponent of 2.82 as verified by the Kolmolgorov-Smirnov one sample goodness of fit test. Scourceforge data is broken into developers and administrators, but when both were used as authors the Lotka distribution exponent of 2.55 produces the lowest error. This would not be significant by the K-S test but the 3.54% maximum error would indicate a fit and calls into question the appropriateness of K-S for large populations of authors.
    Source
    Journal of the American Society for Information Science and technology. 54(2003) no.2, S.169-178
  8. Greenberg, J.: Understanding metadata and metadata scheme (2005) 0.00
    0.0035509837 = product of:
      0.010652951 = sum of:
        0.010652951 = product of:
          0.021305902 = sum of:
            0.021305902 = weight(_text_:of in 5725) [ClassicSimilarity], result of:
              0.021305902 = score(doc=5725,freq=18.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.3109903 = fieldWeight in 5725, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5725)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Although the development and implementation of metadata schemes over the last decade has been extensive, research examining the sum of these activities is limited. This limitation is likely due to the massive scope of the topic. A framework is needed to study the full extent of, and functionalities supported by, metadata schemes. Metadata schemes developed for information resources are analyzed. To begin, I present a review of the definition of metadata, metadata functions, and several metadata typologies. Next, a conceptualization for metadata schemes is presented. The emphasis is on semantic container-like metadata schemes (data structures). The last part of this paper introduces the MODAL (Metadata Objectives and principles, Domains, and Architectural Layout) framework as an approach for studying metadata schemes. The paper concludes with a brief discussion on value of frameworks for examining metadata schemes, including different types of metadata schemes.
  9. Greenberg, J.: Subject control of ephemera : MARC format options (1996) 0.00
    0.003382594 = product of:
      0.010147782 = sum of:
        0.010147782 = product of:
          0.020295564 = sum of:
            0.020295564 = weight(_text_:of in 543) [ClassicSimilarity], result of:
              0.020295564 = score(doc=543,freq=12.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.29624295 = fieldWeight in 543, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=543)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Provides an overview of the MARC format and the structure of the bibliographic MARC record. Discusses the MARC-AMC and MARC-VM formats as options for controlling ephemera, lists popular controlled vocabulary tools for subject control over ephemera material and examines subject analysis methodologies. Considers the specific MARC field options for the subject control of ephemera and provides 3 worked examples. Concludes that, while it can be argued that the MARC format does not provide an ideal control system for ephemera, it does offer an excellent means of controlling ephemera in the online environment and permits ephemera to be intellectually linked with related materials of all formats
  10. Greenberg, J.: Metadata and digital information (2009) 0.00
    0.0033478998 = product of:
      0.010043699 = sum of:
        0.010043699 = product of:
          0.020087399 = sum of:
            0.020087399 = weight(_text_:of in 4697) [ClassicSimilarity], result of:
              0.020087399 = score(doc=4697,freq=16.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.2932045 = fieldWeight in 4697, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.046875 = fieldNorm(doc=4697)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    The range of metadata activity over this last decade is both extensive and astonishing, and substantiates metadata as an integral part of our digital information infrastructure. This entry begins with a brief history of metadata relating to digital information, followed by an overview of different metadata types, functions, and domain-specific definitions. Next, the family of standards comprising a metadata architecture are defined, followed by an overview of metadata generation processes, applications, and people: this latter section gives particular attention to automatic metadata generation approaches. The following section explores four key metadata models. The conclusion summarizes the entry, highlights a number of significant metadata challenges, and notes efforts underway to address metadata challenges in the new millennium.
    Source
    Encyclopedia of library and information sciences. 3rd ed. Ed.: M.J. Bates
  11. Crystal, A.; Greenberg, J.: Relevance criteria identified by health information users during Web searches (2006) 0.00
    0.003271467 = product of:
      0.009814401 = sum of:
        0.009814401 = product of:
          0.019628802 = sum of:
            0.019628802 = weight(_text_:of in 5909) [ClassicSimilarity], result of:
              0.019628802 = score(doc=5909,freq=22.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.28651062 = fieldWeight in 5909, product of:
                  4.690416 = tf(freq=22.0), with freq of:
                    22.0 = termFreq=22.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5909)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    This article focuses on the relevance judgments made by health information users who use the Web. Health information users were conceptualized as motivated information users concerned about how an environmental issue affects their health. Users identified their own environmental health interests and conducted a Web search of a particular environmental health Web site. Users were asked to identify (by highlighting with a mouse) the criteria they use to assess relevance in both Web search engine surrogates and full-text Web documents. Content analysis of document criteria highlighted by users identified the criteria these users relied on most often. Key criteria identified included (in order of frequency of appearance) research, topic, scope, data, influence, affiliation, Web characteristics, and authority/ person. A power-law distribution of criteria was observed (a few criteria represented most of the highlighted regions, with a long tail of occasionally used criteria). Implications of this work are that information retrieval (IR) systems should be tailored in terms of users' tendencies to rely on certain document criteria, and that relevance research should combine methods to gather richer, contextualized data. Metadata for IR systems, such as that used in search engine surrogates, could be improved by taking into account actual usage of relevance criteria. Such metadata should be user-centered (based on data from users, as in this study) and contextappropriate (fit to users' situations and tasks).
    Source
    Journal of the American Society for Information Science and Technology. 57(2006) no.10, S.1368-1382
  12. Greenberg, J.: Intellectual control of visual archives : a comparison between the Art and Architecture Thesaurus and the Library of Congress Thesaurus for Graphic Materials (1993) 0.00
    0.0031316737 = product of:
      0.009395021 = sum of:
        0.009395021 = product of:
          0.018790042 = sum of:
            0.018790042 = weight(_text_:of in 546) [ClassicSimilarity], result of:
              0.018790042 = score(doc=546,freq=14.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.2742677 = fieldWeight in 546, product of:
                  3.7416575 = tf(freq=14.0), with freq of:
                    14.0 = termFreq=14.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.046875 = fieldNorm(doc=546)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    The following investigation is a comparison between the Art and Architecture Thesaurus (AAT) and the LC Thesaurus for Graphic Materials (LCTGM), two popular sources for providing subject access to visual archives. The analysis begins with a discussion on the nature of visual archives and the employment of archival control theory to graphic materials. The major difference observed is that the AAT is a faceted structure geared towards a specialized audience of art and architecture researchers, while LCTGM is similar to LCSH in structure and aims to service the wide-spread archival community. The conclusion recognizes the need to understand the differences between subject thesauri and subject heading lists, and the pressing need to investigate and understand intellectual control of visual archives in today's automated environment.
    Object
    Library of Congress thesaurus for graphical materials
  13. Greenberg, J.: Theoretical considerations of lifecycle modeling : an analysis of the Dryad Repository demonstrating automatic metadata propagation, inheritance, and value system adoption (2009) 0.00
    0.0029591531 = product of:
      0.008877459 = sum of:
        0.008877459 = product of:
          0.017754918 = sum of:
            0.017754918 = weight(_text_:of in 2990) [ClassicSimilarity], result of:
              0.017754918 = score(doc=2990,freq=18.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.25915858 = fieldWeight in 2990, product of:
                  4.2426405 = tf(freq=18.0), with freq of:
                    18.0 = termFreq=18.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=2990)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    The Dryad repository is for data supporting published research in the field of evolutionary biology and related disciplines. Dryad development team members seek a theoretical framework to aid communication about metadata issues and plans. This article explores lifecycle modeling as a theoretical framework for understanding metadata in the repostiroy enivornment. A background discussion reviews the importance of theory, the status of a metadata theory, and lifecycle concepts. An analysis draws examples from the Dryad repository demonstrating automatic propagation, metadata inheritance, and value system adoption, and reports results from a faceted term mapping experiment that included 12 vocabularies and approximately 600 terms. The article also reports selected key findings from a recent survey on the data-sharing attitudes and behaviors of nearly 400 evolutionary biologists. Te results confirm the applicability of lifecycle modeling to Dryad's metadata infrastructure. The article concludes that lifecycle modeling provides a theoretical framework that can enhance our understanding of metadata, aid communication about the topic of metadata in the repository environment, and potentially help sustain robust repository development.
  14. Greenberg, J.; Murillo, A.; Ogletree, A.; Boyles, R.; Martin, N.; Romeo, C.: Metadata capital : automating metadata workflows in the NIEHS viral vector core laboratory (2014) 0.00
    0.0028993662 = product of:
      0.008698098 = sum of:
        0.008698098 = product of:
          0.017396197 = sum of:
            0.017396197 = weight(_text_:of in 1566) [ClassicSimilarity], result of:
              0.017396197 = score(doc=1566,freq=12.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.25392252 = fieldWeight in 1566, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.046875 = fieldNorm(doc=1566)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    This paper presents research examining metadata capital in the context of the Viral Vector Core Laboratory at the National Institute of Environmental Health Sciences (NIEHS). Methods include collaborative workflow modeling and a metadata analysis. Models of the laboratory's workflow and metadata activity are generated to identify potential opportunities for defining microservices that may be supported by iRODS rules. Generic iRODS rules are also shared along with images of the iRODS prototype. The discussion includes an exploration of a modified capital sigma equation to understand metadata as an asset. The work aims to raise awareness of metadata as an asset and to incentivize investment in metadata R&D.
  15. Li, K.; Greenberg, J.; Dunic, J.: Data objects and documenting scientific processes : an analysis of data events in biodiversity data papers (2020) 0.00
    0.0027899165 = product of:
      0.008369749 = sum of:
        0.008369749 = product of:
          0.016739499 = sum of:
            0.016739499 = weight(_text_:of in 5615) [ClassicSimilarity], result of:
              0.016739499 = score(doc=5615,freq=16.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.24433708 = fieldWeight in 5615, product of:
                  4.0 = tf(freq=16.0), with freq of:
                    16.0 = termFreq=16.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5615)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    The data paper, an emerging scholarly genre, describes research data sets and is intended to bridge the gap between the publication of research data and scientific articles. Research examining how data papers report data events, such as data transactions and manipulations, is limited. The research reported on in this article addresses this limitation and investigated how data events are inscribed in data papers. A content analysis was conducted examining the full texts of 82 data papers, drawn from the curated list of data papers connected to the Global Biodiversity Information Facility. Data events recorded for each paper were organized into a set of 17 categories. Many of these categories are described together in the same sentence, which indicates the messiness of data events in the laboratory space. The findings challenge the degrees to which data papers are a distinct genre compared to research articles and they describe data-centric research processes in a through way. This article also discusses how our results could inform a better data publication ecosystem in the future.
    Source
    Journal of the Association for Information Science and Technology. 71(2020) no.2, S.172-182
  16. Greenberg, J.; Zhao, X.; Monselise, M.; Grabus, S.; Boone, J.: Knowledge organization systems : a network for AI with helping interdisciplinary vocabulary engineering (2021) 0.00
    0.0027618767 = product of:
      0.00828563 = sum of:
        0.00828563 = product of:
          0.01657126 = sum of:
            0.01657126 = weight(_text_:of in 719) [ClassicSimilarity], result of:
              0.01657126 = score(doc=719,freq=8.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.24188137 = fieldWeight in 719, product of:
                  2.828427 = tf(freq=8.0), with freq of:
                    8.0 = termFreq=8.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=719)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Knowledge Organization Systems (KOS) as networks of knowledge have the potential to inform AI operations. This paper explores natural language processing and machine learning in the context of KOS and Helping Interdisciplinary Vocabulary Engineering (HIVE) technology. The paper presents three use cases: HIVE and Historical Knowledge Networks, HIVE for Materials Science (HIVE-4-MAT), and Using HIVE to Enhance and Explore Medical Ontologies. The background section reviews AI foundations, while the use cases provide a frame of reference for discussing current progress and implications of connecting KOS to AI in digital resource collections.
  17. Greenberg, J.: Optimal query expansion (QE) processing methods with semantically encoded structured thesaurus terminology (2001) 0.00
    0.0026467475 = product of:
      0.007940242 = sum of:
        0.007940242 = product of:
          0.015880484 = sum of:
            0.015880484 = weight(_text_:of in 5750) [ClassicSimilarity], result of:
              0.015880484 = score(doc=5750,freq=10.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.23179851 = fieldWeight in 5750, product of:
                  3.1622777 = tf(freq=10.0), with freq of:
                    10.0 = termFreq=10.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5750)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    While researchers have explored the value of structured thesauri as controlled vocabularies for general information retrieval (IR) activities, they have not identified the optimal query expansion (QE) processing methods for taking advantage of the semantic encoding underlying the terminology in these tools. The study reported on in this article addresses this question, and examined whether QE via semantically encoded thesauri terminology is more effective in the automatic or interactive processing environment. The research found that, regardless of end-users' retrieval goals, synonyms and partial synonyms (SYNs) and narrower terms (NTs) are generally good candidates for automatic QE and that related (RTs) are better candidates for interactive QE. The study also examined end-users' selection of semantically encoded thesauri terms for interactive QE, and explored how retrieval goals and QE processes may be combined in future thesauri-supported IR systems
    Source
    Journal of the American Society for Information Science and technology. 52(2001) no.6, S.487-498
  18. Greenberg, J.: Metadata and the World Wide Web (2002) 0.00
    0.0024161388 = product of:
      0.007248416 = sum of:
        0.007248416 = product of:
          0.014496832 = sum of:
            0.014496832 = weight(_text_:of in 4264) [ClassicSimilarity], result of:
              0.014496832 = score(doc=4264,freq=12.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.21160212 = fieldWeight in 4264, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4264)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Metadata is of paramount importance for persons, organizations, and endeavors of every dimension that are increasingly turning to the World Wide Web (hereafter referred to as the Web) as a chief conduit for accessing and disseminating information. This is evidenced by the development and implementation of metadata schemas supporting projects ranging from restricted corporate intranets, data warehouses, and consumer-oriented electronic commerce enterprises to freely accessible digital libraries, educational initiatives, virtual museums, and other public Web sites. Today's metadata activities are unprecedented because they extend beyond the traditional library environment in an effort to deal with the Web's exponential growth. This article considers metadata in today's Web environment. The article defines metadata, examines the relationship between metadata and cataloging, provides definitions for key metadata vocabulary terms, and explores the topic of metadata generation. Metadata is an extensive and expanding subject that is prevalent in many environments. For practical reasons, this article has elected to concentrate an the information resource domain, which is defined by electronic textual documents, graphical images, archival materials, museum artifacts, and other objects found in both digital and physical information centers (e.g., libraries, museums, record centers, and archives). To show the extent and larger application of metadata, several examples are also drawn from the data warehouse, electronic commerce, open source, and medical communities.
    Source
    Encyclopedia of library and information science. Vol.72, [=Suppl.35]
  19. Greenberg, J.; Mayer-Patel, K.; Trujillo, S.: YouTube: applying FRBR and exploring the multiple description coding compression model (2012) 0.00
    0.0024161388 = product of:
      0.007248416 = sum of:
        0.007248416 = product of:
          0.014496832 = sum of:
            0.014496832 = weight(_text_:of in 1930) [ClassicSimilarity], result of:
              0.014496832 = score(doc=1930,freq=12.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.21160212 = fieldWeight in 1930, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1930)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Nearly everyone who has searched YouTube for a favorite show, movie, news cast, or other known item, has retrieved multiple videos clips (or segments) that appear to duplicate, overlap, and relate. The work presented in this paper considers this challenge and reports on a study examining the applicability of the Functional Requirements for Bibliographic Records (FRBR) for relating varying renderings of YouTube videos. The paper also introduces the Multiple Description Coding Compression (MDC2) to extend FRBR and address YouTube preservation/storage challenges. The study sample included 20 video segments from YouTube; 10 connected with the event, Small Step for Man (US Astronaut Neil Armstrong's first step on the moon), and 10 with the 1966 classic movie, "Batman: The Movie." The FRBR analysis used a qualitative content analysis, and the MDC2 exploration was pursued via high-level approach of protocol modeling. Results indicate that FRBR is applicable to YouTube, although the analyses required a localization of the Work, Expression, Manifestation, and Item (WEMI) FRBR elements. The MDC2 exploration illustrates an approach for exploring FRBR in the context of other models, and identifies a potential means for addressing YouTube-related preservation/storage challenges.
    Content
    Contribution to a special issue "The FRBR family of conceptual models: toward a linked future"
  20. Greenberg, J.: Reference structures : stagnation, progress, and future challenges (1997) 0.00
    0.0023918552 = product of:
      0.0071755657 = sum of:
        0.0071755657 = product of:
          0.014351131 = sum of:
            0.014351131 = weight(_text_:of in 1103) [ClassicSimilarity], result of:
              0.014351131 = score(doc=1103,freq=6.0), product of:
                0.06850986 = queryWeight, product of:
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.043811057 = queryNorm
                0.20947541 = fieldWeight in 1103, product of:
                  2.4494898 = tf(freq=6.0), with freq of:
                    6.0 = termFreq=6.0
                  1.5637573 = idf(docFreq=25162, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=1103)
          0.5 = coord(1/2)
      0.33333334 = coord(1/3)
    
    Abstract
    Assesses the current state of reference structures in OPACs in a framework defined by stagnation, progress, and future challenges. 'Stagnation' referes to the limited and inconsistent reference structure access provided in current OPACs. 'Progress' refers to improved OPAC reference structure access and reference structure possibilities that extend beyond those commonly represented in existing subject autgority control tools. The progress discussion is supported by a look at professional committee work, data modelling ideas, ontological theory, and one area of linguistic research. The discussion ends with a list of 6 areas needing attention if reference structure access is to be improved in the future OPAC environment