Search (491 results, page 1 of 25)

  • × year_i:[2020 TO 2030}
  1. Rügenhagen, M.; Beck, T.S.; Sartorius, E.J.: Information integrity in the era of Fake News : an experiment using library guidelines to judge information integrity (2020) 0.19
    0.19052733 = product of:
      0.38105467 = sum of:
        0.35207906 = weight(_text_:judge in 1114) [ClassicSimilarity], result of:
          0.35207906 = score(doc=1114,freq=2.0), product of:
            0.5152282 = queryWeight, product of:
              7.731176 = idf(docFreq=52, maxDocs=44421)
              0.06664293 = queryNorm
            0.68334585 = fieldWeight in 1114, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.731176 = idf(docFreq=52, maxDocs=44421)
              0.0625 = fieldNorm(doc=1114)
        0.028975615 = weight(_text_:und in 1114) [ClassicSimilarity], result of:
          0.028975615 = score(doc=1114,freq=2.0), product of:
            0.1478073 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.06664293 = queryNorm
            0.19603643 = fieldWeight in 1114, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0625 = fieldNorm(doc=1114)
      0.5 = coord(2/4)
    
    Source
    Bibliothek: Forschung und Praxis. 44(2020) H.1, S.34-53
  2. Auer, S.; Oelen, A.; Haris, A.M.; Stocker, M.; D'Souza, J.; Farfar, K.E.; Vogt, L.; Prinz, M.; Wiens, V.; Jaradeh, M.Y.: Improving access to scientific literature with knowledge graphs : an experiment using library guidelines to judge information integrity (2020) 0.12
    0.11907958 = product of:
      0.23815916 = sum of:
        0.22004941 = weight(_text_:judge in 1317) [ClassicSimilarity], result of:
          0.22004941 = score(doc=1317,freq=2.0), product of:
            0.5152282 = queryWeight, product of:
              7.731176 = idf(docFreq=52, maxDocs=44421)
              0.06664293 = queryNorm
            0.42709115 = fieldWeight in 1317, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.731176 = idf(docFreq=52, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1317)
        0.01810976 = weight(_text_:und in 1317) [ClassicSimilarity], result of:
          0.01810976 = score(doc=1317,freq=2.0), product of:
            0.1478073 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.06664293 = queryNorm
            0.12252277 = fieldWeight in 1317, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1317)
      0.5 = coord(2/4)
    
    Source
    Bibliothek: Forschung und Praxis. 44(2020) H.3, S.516-529
  3. Ostrzinski, U.: Deutscher MeSH : ZB MED veröffentlicht aktuelle Jahresversion 2022 - freier Zugang und FAIRe Dateiformate (2022) 0.07
    0.06952514 = product of:
      0.13905028 = sum of:
        0.04097771 = weight(_text_:und in 1626) [ClassicSimilarity], result of:
          0.04097771 = score(doc=1626,freq=16.0), product of:
            0.1478073 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.06664293 = queryNorm
            0.2772374 = fieldWeight in 1626, product of:
              4.0 = tf(freq=16.0), with freq of:
                16.0 = termFreq=16.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.03125 = fieldNorm(doc=1626)
        0.09807256 = weight(_text_:headings in 1626) [ClassicSimilarity], result of:
          0.09807256 = score(doc=1626,freq=4.0), product of:
            0.32337824 = queryWeight, product of:
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.06664293 = queryNorm
            0.30327508 = fieldWeight in 1626, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.03125 = fieldNorm(doc=1626)
      0.5 = coord(2/4)
    
    Abstract
    Die aktuelle Ausgabe 2022 der deutschen Ausgabe der Medical Subject Headings (MeSH) steht ab sofort zum Download in verschiedenen FAIRen Dateiformaten sowie als XML- und CSV-Datei unter der CC BY 4.0-Lizenz bereit. Zu den semantisch FAIRen Formaten, die ZB MED anbietet, zählen beispielsweise RDF/XML oder JSON-LD. Sie ermöglichen es etwa Software-Lösungen zur Datenanalyse - auch mit künstlicher Intelligenz -, die Daten direkt zu nutzen. Sie müssen nicht zusätzlich konvertiert und aufbereitet werden.
    Content
    "ZB MED bietet für die deutschsprachigen MeSH-Begriffe einen Internationalized Resource Identifier (IRI) an. Der IRI-Service stellt auf einer HTML-Seite alle Informationen für einen deutschen MeSH-Term bereit und ermöglicht so die Versionierung. Die Sichtbarkeit veralteter, aber in der Vergangenheit genutzter Terme ist im Sinne der FAIR-Prinzipien dadurch weiterhin gewährleistet. Für die Übersetzung der Medical Subject Headings nutzt ZB MED den eigens entwickelten TermCurator. Das semiautomatische Übersetzungstool bietet unter anderem einen integrierten mehrstufigen Kuratierungsprozess. Der MeSH-Thesaurus als polyhierarchisches, konzeptbasiertes Schlagwortregister für biomedizinische Fachbegriffe umfasst das Vokabular, welches in den NLM-Datenbanken, beispielsweise MEDLINE oder PubMed, erscheint. Darüber hinaus ist er eine der wichtigsten Quellen für ein kontrolliertes biomedizinisches Fachvokabular - beispielsweise für die Kategorisierung und Analyse von Literatur- und Datenquellen.
    Der MeSH ist ein polyhierarchisches, konzeptbasiertes Schlagwortregister für biomedizinische Fachbegriffe umfasst das Vokabular, welches in den NLM-Datenbanken, beispielsweise MEDLINE oder PubMed, erscheint. Er wird jährlich aktualisiert von der U. S. National Library of Medicine herausgegeben. Für die deutschsprachige Fassung übersetzt ZB MED dann die jeweils neu hinzugekommenen Terme und ergänzt sie um zusätzliche Synonyme. Erstmalig erstellte ZB MED den Deutschen MeSH im Jahr 2020. Vorher lag die Verantwortung beim Deutschen Institut für Medizinische Dokumentation und Information (DIMDI/BfArM)."
  4. Steeg, F.; Pohl, A.: ¬Ein Protokoll für den Datenabgleich im Web am Beispiel von OpenRefine und der Gemeinsamen Normdatei (GND) (2021) 0.07
    0.067299314 = product of:
      0.13459863 = sum of:
        0.04791392 = weight(_text_:und in 1368) [ClassicSimilarity], result of:
          0.04791392 = score(doc=1368,freq=14.0), product of:
            0.1478073 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.06664293 = queryNorm
            0.32416478 = fieldWeight in 1368, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1368)
        0.08668471 = weight(_text_:headings in 1368) [ClassicSimilarity], result of:
          0.08668471 = score(doc=1368,freq=2.0), product of:
            0.32337824 = queryWeight, product of:
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.06664293 = queryNorm
            0.26805982 = fieldWeight in 1368, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1368)
      0.5 = coord(2/4)
    
    Abstract
    Normdaten spielen speziell im Hinblick auf die Qualität der Inhaltserschließung bibliografischer und archivalischer Ressourcen eine wichtige Rolle. Ein konkretes Ziel der Inhaltserschließung ist z. B., dass alle Werke über Hermann Hesse einheitlich zu finden sind. Hier bieten Normdaten eine Lösung, indem z. B. bei der Erschließung einheitlich die GND-Nummer 11855042X für Hermann Hesse verwendet wird. Das Ergebnis ist eine höhere Qualität der Inhaltserschließung vor allem im Sinne von Einheitlichkeit und Eindeutigkeit und, daraus resultierend, eine bessere Auffindbarkeit. Werden solche Entitäten miteinander verknüpft, z. B. Hermann Hesse mit einem seiner Werke, entsteht ein Knowledge Graph, wie ihn etwa Google bei der Inhaltserschließung des Web verwendet (Singhal 2012). Die Entwicklung des Google Knowledge Graph und das hier vorgestellte Protokoll sind historisch miteinander verbunden: OpenRefine wurde ursprünglich als Google Refine entwickelt, und die Funktionalität zum Abgleich mit externen Datenquellen (Reconciliation) wurde ursprünglich zur Einbindung von Freebase entwickelt, einer der Datenquellen des Google Knowledge Graph. Freebase wurde später in Wikidata integriert. Schon Google Refine wurde zum Abgleich mit Normdaten verwendet, etwa den Library of Congress Subject Headings (Hooland et al. 2013).
    Series
    Bibliotheks- und Informationspraxis; 70
  5. Noever, D.; Ciolino, M.: ¬The Turing deception (2022) 0.07
    0.06601483 = product of:
      0.2640593 = sum of:
        0.2640593 = weight(_text_:judge in 1863) [ClassicSimilarity], result of:
          0.2640593 = score(doc=1863,freq=2.0), product of:
            0.5152282 = queryWeight, product of:
              7.731176 = idf(docFreq=52, maxDocs=44421)
              0.06664293 = queryNorm
            0.5125094 = fieldWeight in 1863, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.731176 = idf(docFreq=52, maxDocs=44421)
              0.046875 = fieldNorm(doc=1863)
      0.25 = coord(1/4)
    
    Abstract
    This research revisits the classic Turing test and compares recent large language models such as ChatGPT for their abilities to reproduce human-level comprehension and compelling text generation. Two task challenges- summary and question answering- prompt ChatGPT to produce original content (98-99%) from a single text entry and sequential questions initially posed by Turing in 1950. We score the original and generated content against the OpenAI GPT-2 Output Detector from 2019, and establish multiple cases where the generated content proves original and undetectable (98%). The question of a machine fooling a human judge recedes in this work relative to the question of "how would one prove it?" The original contribution of the work presents a metric and simple grammatical set for understanding the writing mechanics of chatbots in evaluating their readability and statistical clarity, engagement, delivery, overall quality, and plagiarism risks. While Turing's original prose scores at least 14% below the machine-generated output, whether an algorithm displays hints of Turing's true initial thoughts (the "Lovelace 2.0" test) remains unanswerable.
  6. Lorenzo, L.; Mak, L.; Smeltekop, N.: FAST Headings in MODS : Michigan State University libraries digital repository case study (2023) 0.06
    0.060679298 = product of:
      0.24271719 = sum of:
        0.24271719 = weight(_text_:headings in 2179) [ClassicSimilarity], result of:
          0.24271719 = score(doc=2179,freq=8.0), product of:
            0.32337824 = queryWeight, product of:
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.06664293 = queryNorm
            0.7505675 = fieldWeight in 2179, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.0546875 = fieldNorm(doc=2179)
      0.25 = coord(1/4)
    
    Abstract
    The Michigan State University Libraries (MSUL) digital repository contains numerous collections of openly available material. Since 2016, the digital repository has been using Faceted Application of Subject Terminology (FAST) subject headings as its primary subject vocabulary in order to streamline faceting, display, and search. The MSUL FAST use case presents some challenges that are not addressed by existing MARC-focused FAST tools. This paper will outline the MSUL digital repository team's justification for including FAST headings in the digital repository as well as workflows for adding FAST headings to Metadata Object Description Schema (MODS) metadata, their maintenance, and utilization for discovery.
  7. Zhang, Y.; Trace, C.B.: ¬The quality of health and wellness self-tracking data : a consumer perspective (2022) 0.06
    0.055012353 = product of:
      0.22004941 = sum of:
        0.22004941 = weight(_text_:judge in 1460) [ClassicSimilarity], result of:
          0.22004941 = score(doc=1460,freq=2.0), product of:
            0.5152282 = queryWeight, product of:
              7.731176 = idf(docFreq=52, maxDocs=44421)
              0.06664293 = queryNorm
            0.42709115 = fieldWeight in 1460, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.731176 = idf(docFreq=52, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1460)
      0.25 = coord(1/4)
    
    Abstract
    Information quality (IQ) is key to users' satisfaction with information systems. Understanding what IQ means to users can effectively inform system improvement. Existing inquiries into self-tracking data quality primarily focus on accuracy. Interviewing 20 consumers who had self-tracked health indicators for at least 6 months, we identified eight dimensions that consumers apply to evaluate self-tracking data quality: value-added, accuracy, completeness, accessibility, ease of understanding, trustworthiness, aesthetics, and invasiveness. These dimensions fell into four categories-intrinsic, contextual, representational, and accessibility-suggesting that consumers judge self-tracking data quality not only based on the data's inherent quality but also considering tasks at hand, the clarity of data representation, and data accessibility. We also found that consumers' self-tracking data quality judgments are shaped primarily by their goals or motivations, subjective experience with tracked activities, mental models of how systems work, self-tracking tools' reputation, cost, and design, and domain knowledge and intuition, but less by more objective criteria such as scientific research results, validated devices, or consultation with experts. Future studies should develop and validate a scale for measuring consumers' perceptions of self-tracking data quality and commit efforts to develop technologies and training materials to enhance consumers' ability to evaluate data quality.
  8. Tang, M.-C.; Liao, I.-H.: Preference diversity and openness to novelty : scales construction from the perspective of movie recommendation (2022) 0.06
    0.055012353 = product of:
      0.22004941 = sum of:
        0.22004941 = weight(_text_:judge in 1649) [ClassicSimilarity], result of:
          0.22004941 = score(doc=1649,freq=2.0), product of:
            0.5152282 = queryWeight, product of:
              7.731176 = idf(docFreq=52, maxDocs=44421)
              0.06664293 = queryNorm
            0.42709115 = fieldWeight in 1649, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.731176 = idf(docFreq=52, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1649)
      0.25 = coord(1/4)
    
    Abstract
    In response to calls for recommender systems to balance accuracy and alternative measures such as diversity and novelty, we propose that recommendation strategies should be applied adaptively according to users' preference traits. Psychological scales for "preference diversity" and "openness to novelty" were developed to measure users' willingness to accept diverse and novel recommendations, respectively. To validate the scales empirically, a user study was conducted in which 293 regular moviegoers were asked to judge a set of 220 movies representing both mainstream and "long-tail" appeals. The judgment task involved indicating and rating movies they had seen, heard of but not seen, and not known previously. Correlatoin analyses were then conducted between the participants' preference diversity and openness to novelty scores with the diversity and novelty of their past movie viewing profile and movies they had not seen before but shown interest in. Preference diversity scores were shown to be significantly related to the diversity of the movies they had seen. Higher preference diversity scores were also associated with greater diversity in favored unknown movies. Similarly, participants who scored high on the openness to novelty scale had viewed more little-known movies and were generally interested in less popular movies as well as movies that differed from those they had seen before. Implications of these psychological traits for recommendation strategies are also discussed.
  9. Zimmerman, N.: User study: implementation of OCLC FAST subject headings in the Lafayette digital repository (2023) 0.05
    0.052549813 = product of:
      0.21019925 = sum of:
        0.21019925 = weight(_text_:headings in 2178) [ClassicSimilarity], result of:
          0.21019925 = score(doc=2178,freq=6.0), product of:
            0.32337824 = queryWeight, product of:
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.06664293 = queryNorm
            0.6500105 = fieldWeight in 2178, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.0546875 = fieldNorm(doc=2178)
      0.25 = coord(1/4)
    
    Abstract
    Digital repository migrations present a periodic opportunity to assess metadata quality and to perform strategic enhancements. Lafayette College Libraries implemented OCLC FAST (Faceted Application of Subject Terminology) for its digital image collections as part of a migration from multiple repositories to a single one built on the Samvera Hyrax open-source framework. Application of FAST has normalized subject headings across dissimilar collections in a way that tremendously improves descriptive consistency for staff and discoverability for end users. However, the process of applying FAST headings was complicated by several features of in-scope metadata as well as gaps in available controlled subject authorities.
  10. Hutchinson, J.; Nakatomi, J.: Improving subject description of an LGBTQ+ collection (2024) 0.05
    0.04903628 = product of:
      0.19614512 = sum of:
        0.19614512 = weight(_text_:headings in 2159) [ClassicSimilarity], result of:
          0.19614512 = score(doc=2159,freq=4.0), product of:
            0.32337824 = queryWeight, product of:
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.06664293 = queryNorm
            0.60655016 = fieldWeight in 2159, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.0625 = fieldNorm(doc=2159)
      0.25 = coord(1/4)
    
    Abstract
    This article summarizes the work done as part of a project to improve subject description of an LGBTQ + collection in the ONE Archives, part of the University of Southern California (USC) Libraries. The project involved adding local subject headings to augment existing Library of Congress Subject Headings. The article describes the steps that the project team took, along with the methods that were rejected. The paper discusses reasons why the team chose their course of action.
  11. Smith, A.: Physics Subject Headings (PhySH) (2020) 0.05
    0.0450427 = product of:
      0.1801708 = sum of:
        0.1801708 = weight(_text_:headings in 884) [ClassicSimilarity], result of:
          0.1801708 = score(doc=884,freq=6.0), product of:
            0.32337824 = queryWeight, product of:
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.06664293 = queryNorm
            0.5571519 = fieldWeight in 884, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.046875 = fieldNorm(doc=884)
      0.25 = coord(1/4)
    
    Abstract
    PhySH (Physics Subject Headings) was developed by the American Physical Society and first used in 2016 as a faceted hierarchical controlled vocabulary for physics, with some basic terms from related fields. It was developed mainly for the purpose of associating subjects with papers submitted to and published in the Physical Review family of journals. The scheme is organized at the top level with a two-dimensional classification, with one dimension (labeled "disciplines") representing professional divisions within physics, and the other dimension (labeled "facets") providing a conceptual partitioning of terms. PhySH was preceded in use by PACS ("Physics and Astronomy Classification Scheme"), which was in turn preceded by more ad hoc approaches, and this history and related vocabularies or categorizations will also be briefly discussed.
    Object
    Physics Subject Headings
  12. Bullard, J.; Watson, B.; Purdome, C.: Misrepresentation in the surrogate : author critiques of "Indians of North America" subject headings (2022) 0.04
    0.042906743 = product of:
      0.17162697 = sum of:
        0.17162697 = weight(_text_:headings in 2145) [ClassicSimilarity], result of:
          0.17162697 = score(doc=2145,freq=4.0), product of:
            0.32337824 = queryWeight, product of:
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.06664293 = queryNorm
            0.5307314 = fieldWeight in 2145, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.0546875 = fieldNorm(doc=2145)
      0.25 = coord(1/4)
    
    Abstract
    The surrogate record for a book in the library catalog contains subject headings applied on the basis of literary warrant. To assess the extent to which terms like "Indians of North America" are accurate to the content of the items with that label, we invited the items' creators to critique their surrogate records. In interviews with 38 creators we found consensus against the term "Indians of North America" and identified a periphery of related terms that misrepresent the content of the work, are out of alignment with their scholarly communities, and reproduce settler colonial biases in our library systems.
  13. Cooey, N.; Phillips, A.: Library of Congress Subject Headings : a post-coordinated future (2023) 0.04
    0.042906743 = product of:
      0.17162697 = sum of:
        0.17162697 = weight(_text_:headings in 2165) [ClassicSimilarity], result of:
          0.17162697 = score(doc=2165,freq=4.0), product of:
            0.32337824 = queryWeight, product of:
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.06664293 = queryNorm
            0.5307314 = fieldWeight in 2165, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.0546875 = fieldNorm(doc=2165)
      0.25 = coord(1/4)
    
    Abstract
    This paper is the result of a request from Library of Congress leadership to assess pre-coordinated versus post-coordinated subject cataloging. It argues that the disadvantages of pre-coordinated subject strings are perennial and continue to hinder progress, while the advantages of post-coordinated subject cataloging have expanded, resulting in new opportunities to serve the needs of catalogers and end users alike. The consequences of retaining pre-coordinated headings will have long-term impacts that heavily out-weigh the short-term challenges of transitioning to new cataloging practices. By implementing post-coordinated, faceted vocabularies, the Library of Congress will be investing in the future of libraries.
  14. Walker, J.M.: Faceted vocabularies in catalog searches : provenance evidence vocabulary as search terms or limiters for a personal library collection (2023) 0.04
    0.042906743 = product of:
      0.17162697 = sum of:
        0.17162697 = weight(_text_:headings in 2175) [ClassicSimilarity], result of:
          0.17162697 = score(doc=2175,freq=4.0), product of:
            0.32337824 = queryWeight, product of:
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.06664293 = queryNorm
            0.5307314 = fieldWeight in 2175, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.0546875 = fieldNorm(doc=2175)
      0.25 = coord(1/4)
    
    Abstract
    Genre/Form headings are an important means by which librarians provide users with contextual or descriptive information. To facilitate the discovery of resources with important provenance characteristics, the Marion E. Wade Center added terms from a controlled vocabulary to bibliographic records representing items in the C. S. Lewis personal library collection. The selected terms focus on features that have historically been of interest to visitors. The addition of these headings in the bibliographic records allows users to use these keywords to conduct a search or narrow their results, resulting in more flexibility to locate and select the resources that best meet their needs.
  15. Wlodarczyk, B.: KABA Subject Headings and the National Library of Poland Descriptors in light of Wojciech Wrzosek's theory of historiographical metaphors and different historiographical traditions (2020) 0.04
    0.03753558 = product of:
      0.15014233 = sum of:
        0.15014233 = weight(_text_:headings in 733) [ClassicSimilarity], result of:
          0.15014233 = score(doc=733,freq=6.0), product of:
            0.32337824 = queryWeight, product of:
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.06664293 = queryNorm
            0.46429324 = fieldWeight in 733, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.0390625 = fieldNorm(doc=733)
      0.25 = coord(1/4)
    
    Abstract
    The aims of this article are, first, to provide a necessary background to investigate the discipline of history from the knowledge organization (KO) perspective, and econdly, to present, on selected examples, a way of analyzing knowledge organization systems (KOSs) from the point of view of the theory of history. The study includes a literature review and epistemological analysis. It provides a preliminary analysis of history in two selected universal Polish KOSs: KABA subject headings and the National Library of Poland Descriptors. The research is restricted to the high-level concept of historiographical metaphors coined by Wojciech Wrzosek and how they can be utilized in analyzing KOSs. The analysis of the structure of the KOSs and indexing practices of selected history books is performed. A particular emphasis is placed upon the requirements of classical and non-classical historiography in the context of KO. Although the knowledge about historiographical metaphors given by Wrzosek can be helpful for the analysis and improvement of KOSs, it seems that their broad character can provide the creators only with some general guidelines. Historical research is multidimensional, which is why the general remarks presented in this article need to be supplemented with in-depth theoretical and empirical analyses of historiography.
    Object
    KABA Subject Headings
  16. Peponakis, M.; Mastora, A.; Kapidakis, S.; Doerr, M.: Expressiveness and machine processability of Knowledge Organization Systems (KOS) : an analysis of concepts and relations (2020) 0.03
    0.030647675 = product of:
      0.1225907 = sum of:
        0.1225907 = weight(_text_:headings in 787) [ClassicSimilarity], result of:
          0.1225907 = score(doc=787,freq=4.0), product of:
            0.32337824 = queryWeight, product of:
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.06664293 = queryNorm
            0.37909386 = fieldWeight in 787, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.0390625 = fieldNorm(doc=787)
      0.25 = coord(1/4)
    
    Abstract
    This study considers the expressiveness (that is the expressive power or expressivity) of different types of Knowledge Organization Systems (KOS) and discusses its potential to be machine-processable in the context of the Semantic Web. For this purpose, the theoretical foundations of KOS are reviewed based on conceptualizations introduced by the Functional Requirements for Subject Authority Data (FRSAD) and the Simple Knowledge Organization System (SKOS); natural language processing techniques are also implemented. Applying a comparative analysis, the dataset comprises a thesaurus (Eurovoc), a subject headings system (LCSH) and a classification scheme (DDC). These are compared with an ontology (CIDOC-CRM) by focusing on how they define and handle concepts and relations. It was observed that LCSH and DDC focus on the formalism of character strings (nomens) rather than on the modelling of semantics; their definition of what constitutes a concept is quite fuzzy, and they comprise a large number of complex concepts. By contrast, thesauri have a coherent definition of what constitutes a concept, and apply a systematic approach to the modelling of relations. Ontologies explicitly define diverse types of relations, and are by their nature machine-processable. The paper concludes that the potential of both the expressiveness and machine processability of each KOS is extensively regulated by its structural rules. It is harder to represent subject headings and classification schemes as semantic networks with nodes and arcs, while thesauri are more suitable for such a representation. In addition, a paradigm shift is revealed which focuses on the modelling of relations between concepts, rather than the concepts themselves.
  17. Ahmed, M.; Mukhopadhyay, M.; Mukhopadhyay, P.: Automated knowledge organization : AI ML based subject indexing system for libraries (2023) 0.03
    0.030647675 = product of:
      0.1225907 = sum of:
        0.1225907 = weight(_text_:headings in 1979) [ClassicSimilarity], result of:
          0.1225907 = score(doc=1979,freq=4.0), product of:
            0.32337824 = queryWeight, product of:
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.06664293 = queryNorm
            0.37909386 = fieldWeight in 1979, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1979)
      0.25 = coord(1/4)
    
    Abstract
    The research study as reported here is an attempt to explore the possibilities of an AI/ML-based semi-automated indexing system in a library setup to handle large volumes of documents. It uses the Python virtual environment to install and configure an open source AI environment (named Annif) to feed the LOD (Linked Open Data) dataset of Library of Congress Subject Headings (LCSH) as a standard KOS (Knowledge Organisation System). The framework deployed the Turtle format of LCSH after cleaning the file with Skosify, applied an array of backend algorithms (namely TF-IDF, Omikuji, and NN-Ensemble) to measure relative performance, and selected Snowball as an analyser. The training of Annif was conducted with a large set of bibliographic records populated with subject descriptors (MARC tag 650$a) and indexed by trained LIS professionals. The training dataset is first treated with MarcEdit to export it in a format suitable for OpenRefine, and then in OpenRefine it undergoes many steps to produce a bibliographic record set suitable to train Annif. The framework, after training, has been tested with a bibliographic dataset to measure indexing efficiencies, and finally, the automated indexing framework is integrated with data wrangling software (OpenRefine) to produce suggested headings on a mass scale. The entire framework is based on open-source software, open datasets, and open standards.
  18. Grabus, S.; Logan, P.M.; Greenberg, J.: Temporal concept drift and alignment : an empirical approach to comparing knowledge organization systems over time (2022) 0.03
    0.030647675 = product of:
      0.1225907 = sum of:
        0.1225907 = weight(_text_:headings in 2102) [ClassicSimilarity], result of:
          0.1225907 = score(doc=2102,freq=4.0), product of:
            0.32337824 = queryWeight, product of:
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.06664293 = queryNorm
            0.37909386 = fieldWeight in 2102, product of:
              2.0 = tf(freq=4.0), with freq of:
                4.0 = termFreq=4.0
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.0390625 = fieldNorm(doc=2102)
      0.25 = coord(1/4)
    
    Abstract
    This research explores temporal concept drift and temporal alignment in knowledge organization systems (KOS). A comparative analysis is pursued using the 1910 Library of Congress Subject Headings, 2020 FAST Topical, and automatic indexing. The use case involves a sample of 90 nineteenth-century Encyclopedia Britannica entries. The entries were indexed using two approaches: 1) full-text indexing; 2) Named Entity Recognition was performed upon the entries with Stanza, Stanford's NLP toolkit, and entities were automatically indexed with the Helping Interdisciplinary Vocabulary application (HIVE), using both 1910 LCSH and FAST Topical. The analysis focused on three goals: 1) identifying results that were exclusive to the 1910 LCSH output; 2) identifying terms in the exclusive set that have been deprecated from the contemporary LCSH, demonstrating temporal concept drift; and 3) exploring the historical significance of these deprecated terms. Results confirm that historical vocabularies can be used to generate anachronistic subject headings representing conceptual drift across time in KOS and historical resources. A methodological contribution is made demonstrating how to study changes in KOS over time and improve the contextualization historical humanities resources.
  19. Smith, C.: Controlled vocabularies : past, present and future of subject access (2021) 0.03
    0.030339649 = product of:
      0.121358596 = sum of:
        0.121358596 = weight(_text_:headings in 1705) [ClassicSimilarity], result of:
          0.121358596 = score(doc=1705,freq=2.0), product of:
            0.32337824 = queryWeight, product of:
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.06664293 = queryNorm
            0.37528375 = fieldWeight in 1705, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.0546875 = fieldNorm(doc=1705)
      0.25 = coord(1/4)
    
    Abstract
    Controlled vocabularies are a foundational concept in library science and provide a framework for consistency in cataloging practices. Subject headings provide valuable access points to library resources during search and discovery for patrons. Many librarians will be familiar with the more widely used controlled vocabularies, like those maintained by national libraries or major professional organizations. More recently, there has been an increasing shift toward specialized vocabularies maintained by independent entities intended for much narrower use. While there is valid criticism of the nature or content of controlled vocabularies, they will likely continue to be an important feature in information organization.
  20. Moulaison-Sandy, H.; Adkins, D.; Bossaller, J.; Cho, H.: ¬An automated approach to describing fiction : a methodology to use book reviews to identify affect (2021) 0.03
    0.030339649 = product of:
      0.121358596 = sum of:
        0.121358596 = weight(_text_:headings in 1711) [ClassicSimilarity], result of:
          0.121358596 = score(doc=1711,freq=2.0), product of:
            0.32337824 = queryWeight, product of:
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.06664293 = queryNorm
            0.37528375 = fieldWeight in 1711, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              4.8524013 = idf(docFreq=942, maxDocs=44421)
              0.0546875 = fieldNorm(doc=1711)
      0.25 = coord(1/4)
    
    Abstract
    Subject headings and genre terms are notoriously difficult to apply, yet are important for fiction. The current project functions as a proof of concept, using a text-mining methodology to identify affective information (emotion and tone) about fiction titles from professional book reviews as a potential first step in automating the subject analysis process. Findings are presented and discussed, comparing results to the range of aboutness and isness information in library cataloging records. The methodology is likewise presented, and how future work might expand on the current project to enhance catalog records through text-mining is explored.

Languages

  • d 434
  • e 52
  • m 2
  • pt 2
  • More… Less…

Types

  • a 361
  • el 230
  • m 60
  • s 7
  • x 5
  • p 3
  • r 1
  • More… Less…

Subjects

Classifications