Search (171 results, page 1 of 9)

  • × theme_ss:"Wissensrepräsentation"
  1. Bandholtz, T.; Schulte-Coerne, T.; Glaser, R.; Fock, J.; Keller, T.: iQvoc - open source SKOS(XL) maintenance and publishing tool (2010) 0.14
    0.1389289 = product of:
      0.2778578 = sum of:
        0.25281855 = weight(_text_:java in 1604) [ClassicSimilarity], result of:
          0.25281855 = score(doc=1604,freq=2.0), product of:
            0.46384227 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0658165 = queryNorm
            0.5450528 = fieldWeight in 1604, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0546875 = fieldNorm(doc=1604)
        0.025039254 = weight(_text_:und in 1604) [ClassicSimilarity], result of:
          0.025039254 = score(doc=1604,freq=2.0), product of:
            0.14597435 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0658165 = queryNorm
            0.17153187 = fieldWeight in 1604, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0546875 = fieldNorm(doc=1604)
      0.5 = coord(2/4)
    
    Abstract
    iQvoc is a new open source SKOS-XL vocabulary management tool developed by the Federal Environment Agency, Germany, and innoQ Deutschland GmbH. Its immediate purpose is maintaining and publishing reference vocabularies in the upcoming Linked Data cloud of environmental information, but it may be easily adapted to host any SKOS- XL compliant vocabulary. iQvoc is implemented as a Ruby on Rails application running on top of JRuby - the Java implementation of the Ruby Programming Language. To increase the user experience when editing content, iQvoc uses heavily the JavaScript library jQuery.
    Theme
    Konzeption und Anwendung des Prinzips Thesaurus
  2. Nix, M.: ¬Die praktische Einsetzbarkeit des CIDOC CRM in Informationssystemen im Bereich des Kulturerbes (2004) 0.11
    0.10817753 = product of:
      0.21635506 = sum of:
        0.18058468 = weight(_text_:java in 729) [ClassicSimilarity], result of:
          0.18058468 = score(doc=729,freq=2.0), product of:
            0.46384227 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0658165 = queryNorm
            0.38932347 = fieldWeight in 729, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0390625 = fieldNorm(doc=729)
        0.035770364 = weight(_text_:und in 729) [ClassicSimilarity], result of:
          0.035770364 = score(doc=729,freq=8.0), product of:
            0.14597435 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0658165 = queryNorm
            0.24504554 = fieldWeight in 729, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0390625 = fieldNorm(doc=729)
      0.5 = coord(2/4)
    
    Abstract
    Es steht uns eine praktisch unbegrenzte Menge an Informationen über das World Wide Web zur Verfügung. Das Problem, das daraus erwächst, ist, diese Menge zu bewältigen und an die Information zu gelangen, die im Augenblick benötigt wird. Das überwältigende Angebot zwingt sowohl professionelle Anwender als auch Laien zu suchen, ungeachtet ihrer Ansprüche an die gewünschten Informationen. Um dieses Suchen effizienter zu gestalten, gibt es einerseits die Möglichkeit, leistungsstärkere Suchmaschinen zu entwickeln. Eine andere Möglichkeit ist, Daten besser zu strukturieren, um an die darin enthaltenen Informationen zu gelangen. Hoch strukturierte Daten sind maschinell verarbeitbar, sodass ein Teil der Sucharbeit automatisiert werden kann. Das Semantic Web ist die Vision eines weiterentwickelten World Wide Web, in dem derart strukturierten Daten von so genannten Softwareagenten verarbeitet werden. Die fortschreitende inhaltliche Strukturierung von Daten wird Semantisierung genannt. Im ersten Teil der Arbeit sollen einige wichtige Methoden der inhaltlichen Strukturierung von Daten skizziert werden, um die Stellung von Ontologien innerhalb der Semantisierung zu klären. Im dritten Kapitel wird der Aufbau und die Aufgabe des CIDOC Conceptual Reference Model (CRM), einer Domain Ontologie im Bereich des Kulturerbes dargestellt. Im darauf folgenden praktischen Teil werden verschiedene Ansätze zur Verwendung des CRM diskutiert und umgesetzt. Es wird ein Vorschlag zur Implementierung des Modells in XML erarbeitet. Das ist eine Möglichkeit, die dem Datentransport dient. Außerdem wird der Entwurf einer Klassenbibliothek in Java dargelegt, auf die die Verarbeitung und Nutzung des Modells innerhalb eines Informationssystems aufbauen kann.
  3. Botana Varela, J.: Unscharfe Wissensrepräsentationen bei der Implementation des Semantic Web (2004) 0.10
    0.09900196 = product of:
      0.19800392 = sum of:
        0.14446774 = weight(_text_:java in 346) [ClassicSimilarity], result of:
          0.14446774 = score(doc=346,freq=2.0), product of:
            0.46384227 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0658165 = queryNorm
            0.31145877 = fieldWeight in 346, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.03125 = fieldNorm(doc=346)
        0.053536177 = weight(_text_:und in 346) [ClassicSimilarity], result of:
          0.053536177 = score(doc=346,freq=28.0), product of:
            0.14597435 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0658165 = queryNorm
            0.36675057 = fieldWeight in 346, product of:
              5.2915025 = tf(freq=28.0), with freq of:
                28.0 = termFreq=28.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.03125 = fieldNorm(doc=346)
      0.5 = coord(2/4)
    
    Abstract
    In der vorliegenden Arbeit soll einen Ansatz zur Implementation einer Wissensrepräsentation mit den in Abschnitt 1.1. skizzierten Eigenschaften und dem Semantic Web als Anwendungsbereich vorgestellt werden. Die Arbeit ist im Wesentlichen in zwei Bereiche gegliedert: dem Untersuchungsbereich (Kapitel 2-5), in dem ich die in Abschnitt 1.1. eingeführte Terminologie definiert und ein umfassender Überblick über die zugrundeliegenden Konzepte gegeben werden soll, und dem Implementationsbereich (Kapitel 6), in dem aufbauend auf dem im Untersuchungsbereich erarbeiteten Wissen einen semantischen Suchdienst entwickeln werden soll. In Kapitel 2 soll zunächst das Konzept der semantischen Interpretation erläutert und in diesem Kontext hauptsächlich zwischen Daten, Information und Wissen unterschieden werden. In Kapitel 3 soll Wissensrepräsentation aus einer kognitiven Perspektive betrachtet und in diesem Zusammenhang das Konzept der Unschärfe beschrieben werden. In Kapitel 4 sollen sowohl aus historischer als auch aktueller Sicht die Ansätze zur Wissensrepräsentation und -auffindung beschrieben und in diesem Zusammenhang das Konzept der Unschärfe diskutiert werden. In Kapitel 5 sollen die aktuell im WWW eingesetzten Modelle und deren Einschränkungen erläutert werden. Anschließend sollen im Kontext der Entscheidungsfindung die Anforderungen beschrieben werden, die das WWW an eine adäquate Wissensrepräsentation stellt, und anhand der Technologien des Semantic Web die Repräsentationsparadigmen erläutert werden, die diese Anforderungen erfüllen. Schließlich soll das Topic Map-Paradigma erläutert werden. In Kapitel 6 soll aufbauend auf die im Untersuchtungsbereich gewonnenen Erkenntnisse ein Prototyp entwickelt werden. Dieser besteht im Wesentlichen aus Softwarewerkzeugen, die das automatisierte und computergestützte Extrahieren von Informationen, das unscharfe Modellieren, sowie das Auffinden von Wissen unterstützen. Die Implementation der Werkzeuge erfolgt in der Programmiersprache Java, und zur unscharfen Wissensrepräsentation werden Topic Maps eingesetzt. Die Implementation wird dabei schrittweise vorgestellt. Schließlich soll der Prototyp evaluiert und ein Ausblick auf zukünftige Erweiterungsmöglichkeiten gegeben werden. Und schließlich soll in Kapitel 7 eine Synthese formuliert werden.
  4. Lukasiewicz, T.: Uncertainty reasoning for the Semantic Web (2017) 0.09
    0.08670966 = product of:
      0.34683865 = sum of:
        0.34683865 = weight(_text_:handling in 4939) [ClassicSimilarity], result of:
          0.34683865 = score(doc=4939,freq=6.0), product of:
            0.4128091 = queryWeight, product of:
              6.272122 = idf(docFreq=227, maxDocs=44421)
              0.0658165 = queryNorm
            0.84019136 = fieldWeight in 4939, product of:
              2.4494898 = tf(freq=6.0), with freq of:
                6.0 = termFreq=6.0
              6.272122 = idf(docFreq=227, maxDocs=44421)
              0.0546875 = fieldNorm(doc=4939)
      0.25 = coord(1/4)
    
    Abstract
    The Semantic Web has attracted much attention, both from academia and industry. An important role in research towards the Semantic Web is played by formalisms and technologies for handling uncertainty and/or vagueness. In this paper, I first provide some motivating examples for handling uncertainty and/or vagueness in the Semantic Web. I then give an overview of some own formalisms for handling uncertainty and/or vagueness in the Semantic Web.
  5. Resource Description Framework (RDF) : Concepts and Abstract Syntax (2004) 0.06
    0.057213537 = product of:
      0.22885415 = sum of:
        0.22885415 = weight(_text_:handling in 54) [ClassicSimilarity], result of:
          0.22885415 = score(doc=54,freq=2.0), product of:
            0.4128091 = queryWeight, product of:
              6.272122 = idf(docFreq=227, maxDocs=44421)
              0.0658165 = queryNorm
            0.5543825 = fieldWeight in 54, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.272122 = idf(docFreq=227, maxDocs=44421)
              0.0625 = fieldNorm(doc=54)
      0.25 = coord(1/4)
    
    Abstract
    The Resource Description Framework (RDF) is a framework for representing information in the Web. RDF Concepts and Abstract Syntax defines an abstract syntax on which RDF is based, and which serves to link its concrete syntax to its formal semantics. It also includes discussion of design goals, key concepts, datatyping, character normalization and handling of URI references.
  6. Tang, X.-B.; Wei Wei, G,-C.L.; Zhu, J.: ¬An inference model of medical insurance fraud detection : based on ontology and SWRL (2017) 0.05
    0.054175403 = product of:
      0.21670161 = sum of:
        0.21670161 = weight(_text_:java in 4615) [ClassicSimilarity], result of:
          0.21670161 = score(doc=4615,freq=2.0), product of:
            0.46384227 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0658165 = queryNorm
            0.46718815 = fieldWeight in 4615, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=4615)
      0.25 = coord(1/4)
    
    Abstract
    Medical insurance fraud is common in many countries' medical insurance systems and represents a serious threat to the insurance funds and the benefits of patients. In this paper, we present an inference model of medical insurance fraud detection, based on a medical detection domain ontology that incorporates the knowledge base provided by the Medical Terminology, NKIMed, and Chinese Library Classification systems. Through analyzing the behaviors of irregular and fraudulent medical services, we defined the scope of the medical domain ontology relevant to the task and built the ontology about medical sciences and medical service behaviors. The ontology then utilizes Semantic Web Rule Language (SWRL) and Java Expert System Shell (JESS) to detect medical irregularities and mine implicit knowledge. The system can be used to improve the management of medical insurance risks.
  7. Frey, J.; Streitmatter, D.; Götz, F.; Hellmann, S.; Arndt, N.: DBpedia Archivo : a Web-Scale interface for ontology archiving under consumer-oriented aspects (2020) 0.05
    0.050061844 = product of:
      0.20024738 = sum of:
        0.20024738 = weight(_text_:handling in 1053) [ClassicSimilarity], result of:
          0.20024738 = score(doc=1053,freq=2.0), product of:
            0.4128091 = queryWeight, product of:
              6.272122 = idf(docFreq=227, maxDocs=44421)
              0.0658165 = queryNorm
            0.48508468 = fieldWeight in 1053, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.272122 = idf(docFreq=227, maxDocs=44421)
              0.0546875 = fieldNorm(doc=1053)
      0.25 = coord(1/4)
    
    Abstract
    While thousands of ontologies exist on the web, a unified sys-tem for handling online ontologies - in particular with respect to discov-ery, versioning, access, quality-control, mappings - has not yet surfacedand users of ontologies struggle with many challenges. In this paper, wepresent an online ontology interface and augmented archive called DB-pedia Archivo, that discovers, crawls, versions and archives ontologies onthe DBpedia Databus. Based on this versioned crawl, different features,quality measures and, if possible, fixes are deployed to handle and sta-bilize the changes in the found ontologies at web-scale. A comparison toexisting approaches and ontology repositories is given.
  8. Wu, D.; Shi, J.: Classical music recording ontology used in a library catalog (2016) 0.05
    0.04514617 = product of:
      0.18058468 = sum of:
        0.18058468 = weight(_text_:java in 4179) [ClassicSimilarity], result of:
          0.18058468 = score(doc=4179,freq=2.0), product of:
            0.46384227 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0658165 = queryNorm
            0.38932347 = fieldWeight in 4179, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0390625 = fieldNorm(doc=4179)
      0.25 = coord(1/4)
    
    Abstract
    In order to improve the organization of classical music information resources, we constructed a classical music recording ontology, on top of which we then designed an online classical music catalog. Our construction of the classical music recording ontology consisted of three steps: identifying the purpose, analyzing the ontology, and encoding the ontology. We identified the main classes and properties of the domain by investigating classical music recording resources and users' information needs. We implemented the ontology in the Web Ontology Language (OWL) using five steps: transforming the properties, encoding the transformed properties, defining ranges of the properties, constructing individuals, and standardizing the ontology. In constructing the online catalog, we first designed the structure and functions of the catalog based on investigations into users' information needs and information-seeking behaviors. Then we extracted classes and properties of the ontology using the Apache Jena application programming interface (API), and constructed a catalog in the Java environment. The catalog provides a hierarchical main page (built using the Functional Requirements for Bibliographic Records (FRBR) model), a classical music information network and integrated information service; this combination of features greatly eases the task of finding classical music recordings and more information about classical music.
  9. Kruk, S.R.; McDaniel, B.: Goals of semantic digital libraries (2009) 0.04
    0.04291015 = product of:
      0.1716406 = sum of:
        0.1716406 = weight(_text_:handling in 365) [ClassicSimilarity], result of:
          0.1716406 = score(doc=365,freq=2.0), product of:
            0.4128091 = queryWeight, product of:
              6.272122 = idf(docFreq=227, maxDocs=44421)
              0.0658165 = queryNorm
            0.41578686 = fieldWeight in 365, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.272122 = idf(docFreq=227, maxDocs=44421)
              0.046875 = fieldNorm(doc=365)
      0.25 = coord(1/4)
    
    Abstract
    Digital libraries have become commodity in the current world of Internet. More and more information is produced, and more and more non-digital information is being rendered available. The new, more user friendly, community-oriented technologies used throughout the Internet are raising the bar of expectations. Digital libraries cannot stand still with their technologies; if not for the sake of handling rapidly growing amount and diversity of information, they must provide for better user experience matching and overgrowing standards set by the industry. The next generation of digital libraries combine technological solutions, such as P2P, SOA, or Grid, with recent research on semantics and social networks. These solutions are put into practice to answer a variety of requirements imposed on digital libraries.
  10. Padmavathi, T.; Krishnamurthy, M.: Ontological representation of knowledge for developing information services in food science and technology (2012) 0.04
    0.04291015 = product of:
      0.1716406 = sum of:
        0.1716406 = weight(_text_:handling in 1839) [ClassicSimilarity], result of:
          0.1716406 = score(doc=1839,freq=2.0), product of:
            0.4128091 = queryWeight, product of:
              6.272122 = idf(docFreq=227, maxDocs=44421)
              0.0658165 = queryNorm
            0.41578686 = fieldWeight in 1839, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.272122 = idf(docFreq=227, maxDocs=44421)
              0.046875 = fieldNorm(doc=1839)
      0.25 = coord(1/4)
    
    Abstract
    Knowledge explosion in various fields during recent years has resulted in the creation of vast amounts of on-line scientific literature. Food Science &Technology (FST) is also an important subject domain where rapid developments are taking place due to diverse research and development activities. As a result, information storage and retrieval has become very complex and current information retrieval systems (IRs) are being challenged in terms of both adequate precision and response time. To overcome these limitations as well as to provide naturallanguage based effective retrieval, a suitable knowledge engineering framework needs to be applied to represent, share and discover information. Semantic web technologies provide mechanisms for creating knowledge bases, ontologies and rules for handling data that promise to improve the quality of information retrieval. Ontologies are the backbone of such knowledge systems. This paper presents a framework for semantic representation of a large repository of content in the domain of FST.
  11. Vlachidis, A.; Binding, C.; Tudhope, D.; May, K.: Excavating grey literature : a case study on the rich indexing of archaeological documents via natural language-processing techniques and knowledge-based resources (2010) 0.04
    0.036116935 = product of:
      0.14446774 = sum of:
        0.14446774 = weight(_text_:java in 935) [ClassicSimilarity], result of:
          0.14446774 = score(doc=935,freq=2.0), product of:
            0.46384227 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0658165 = queryNorm
            0.31145877 = fieldWeight in 935, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.03125 = fieldNorm(doc=935)
      0.25 = coord(1/4)
    
    Abstract
    Purpose - This paper sets out to discuss the use of information extraction (IE), a natural language-processing (NLP) technique to assist "rich" semantic indexing of diverse archaeological text resources. The focus of the research is to direct a semantic-aware "rich" indexing of diverse natural language resources with properties capable of satisfying information retrieval from online publications and datasets associated with the Semantic Technologies for Archaeological Resources (STAR) project. Design/methodology/approach - The paper proposes use of the English Heritage extension (CRM-EH) of the standard core ontology in cultural heritage, CIDOC CRM, and exploitation of domain thesauri resources for driving and enhancing an Ontology-Oriented Information Extraction process. The process of semantic indexing is based on a rule-based Information Extraction technique, which is facilitated by the General Architecture of Text Engineering (GATE) toolkit and expressed by Java Annotation Pattern Engine (JAPE) rules. Findings - Initial results suggest that the combination of information extraction with knowledge resources and standard conceptual models is capable of supporting semantic-aware term indexing. Additional efforts are required for further exploitation of the technique and adoption of formal evaluation methods for assessing the performance of the method in measurable terms. Originality/value - The value of the paper lies in the semantic indexing of 535 unpublished online documents often referred to as "Grey Literature", from the Archaeological Data Service OASIS corpus (Online AccesS to the Index of archaeological investigationS), with respect to the CRM ontological concepts E49.Time Appellation and P19.Physical Object.
  12. OWLED 2009; OWL: Experiences and Directions, Sixth International Workshop, Chantilly, Virginia, USA, 23-24 October 2009, Co-located with ISWC 2009. (2009) 0.03
    0.027087701 = product of:
      0.108350806 = sum of:
        0.108350806 = weight(_text_:java in 378) [ClassicSimilarity], result of:
          0.108350806 = score(doc=378,freq=2.0), product of:
            0.46384227 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0658165 = queryNorm
            0.23359407 = fieldWeight in 378, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0234375 = fieldNorm(doc=378)
      0.25 = coord(1/4)
    
    Content
    Long Papers * Suggestions for OWL 3, Pascal Hitzler. * BestMap: Context-Aware SKOS Vocabulary Mappings in OWL 2, Rinke Hoekstra. * Mechanisms for Importing Modules, Bijan Parsia, Ulrike Sattler and Thomas Schneider. * A Syntax for Rules in OWL 2, Birte Glimm, Matthew Horridge, Bijan Parsia and Peter Patel-Schneider. * PelletSpatial: A Hybrid RCC-8 and RDF/OWL Reasoning and Query Engine, Markus Stocker and Evren Sirin. * The OWL API: A Java API for Working with OWL 2 Ontologies, Matthew Horridge and Sean Bechhofer. * From Justifications to Proofs for Entailments in OWL, Matthew Horridge, Bijan Parsia and Ulrike Sattler. * A Solution for the Man-Man Problem in the Family History Knowledge Base, Dmitry Tsarkov, Ulrike Sattler and Robert Stevens. * Towards Integrity Constraints in OWL, Evren Sirin and Jiao Tao. * Processing OWL2 ontologies using Thea: An application of logic programming, Vangelis Vassiliadis, Jan Wielemaker and Chris Mungall. * Reasoning in Metamodeling Enabled Ontologies, Nophadol Jekjantuk, Gerd Gröner and Jeff Z. Pan.
  13. Wildgen, W.: Semantischer Realismus und Antirealismus in der Sprachtheorie (1992) 0.02
    0.02399549 = product of:
      0.09598196 = sum of:
        0.09598196 = weight(_text_:und in 2139) [ClassicSimilarity], result of:
          0.09598196 = score(doc=2139,freq=10.0), product of:
            0.14597435 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0658165 = queryNorm
            0.6575262 = fieldWeight in 2139, product of:
              3.1622777 = tf(freq=10.0), with freq of:
                10.0 = termFreq=10.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.09375 = fieldNorm(doc=2139)
      0.25 = coord(1/4)
    
    Series
    Philosophie und Geschichte der Wissenschaften; Bd.18
    Source
    Wirklichkeit und Wissen: Realismus, Antirealismus und Wirklichkeits-Konzeptionen in Philosophie und Wissenschaften. Hrsg.: H.J. Sandkühler
  14. Sandkühler, H.J.: Epistemologischer Realismus und die Wirklichkeit des Wissens : eine Verteidigung der Philosophie des Geistes gegen Naturalismus und Reduktionismus (1992) 0.02
    0.021904785 = product of:
      0.08761914 = sum of:
        0.08761914 = weight(_text_:und in 1731) [ClassicSimilarity], result of:
          0.08761914 = score(doc=1731,freq=12.0), product of:
            0.14597435 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0658165 = queryNorm
            0.60023654 = fieldWeight in 1731, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.078125 = fieldNorm(doc=1731)
      0.25 = coord(1/4)
    
    Series
    Philosophie und Geschichte der Wissenschaften; Bd.18
    Source
    Wirklichkeit und Wissen: Realismus, Antirealismus und Wirklichkeits-Konzeptionen in Philosophie und Wissenschaften. Hrsg.: H.J. Sandkühler
  15. Roth, G.; Schwegler, H.: Kognitive Referenz und Selbstreferentialität des Gehirns : ein Beitrag zur Klärung des Verhältnisses zwischen Erkenntnistheorie und Hirnforschung (1992) 0.02
    0.021904785 = product of:
      0.08761914 = sum of:
        0.08761914 = weight(_text_:und in 607) [ClassicSimilarity], result of:
          0.08761914 = score(doc=607,freq=12.0), product of:
            0.14597435 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0658165 = queryNorm
            0.60023654 = fieldWeight in 607, product of:
              3.4641016 = tf(freq=12.0), with freq of:
                12.0 = termFreq=12.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.078125 = fieldNorm(doc=607)
      0.25 = coord(1/4)
    
    Series
    Philosophie und Geschichte der Wissenschaften; Bd.18
    Source
    Wirklichkeit und Wissen: Realismus, Antirealismus und Wirklichkeits-Konzeptionen in Philosophie und Wissenschaften. Hrsg.: H.J. Sandkühler
  16. Kutschera, F. von: ¬Der erkenntnistheoretische Realismus (1992) 0.02
    0.021462217 = product of:
      0.08584887 = sum of:
        0.08584887 = weight(_text_:und in 608) [ClassicSimilarity], result of:
          0.08584887 = score(doc=608,freq=8.0), product of:
            0.14597435 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0658165 = queryNorm
            0.58810925 = fieldWeight in 608, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.09375 = fieldNorm(doc=608)
      0.25 = coord(1/4)
    
    Series
    Philosophie und Geschichte der Wissenschaften; Bd.18
    Source
    Wirklichkeit und Wissen: Realismus, Antirealismus und Wirklichkeits-Konzeptionen in Philosophie und Wissenschaften. Hrsg.: H.J. Sandkühler
  17. Franzen, W.: Idealismus statt Realismus? : Realismus plus Skeptizismus! (1992) 0.02
    0.021462217 = product of:
      0.08584887 = sum of:
        0.08584887 = weight(_text_:und in 612) [ClassicSimilarity], result of:
          0.08584887 = score(doc=612,freq=8.0), product of:
            0.14597435 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0658165 = queryNorm
            0.58810925 = fieldWeight in 612, product of:
              2.828427 = tf(freq=8.0), with freq of:
                8.0 = termFreq=8.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.09375 = fieldNorm(doc=612)
      0.25 = coord(1/4)
    
    Series
    Philosophie und Geschichte der Wissenschaften; Bd.18
    Source
    Wirklichkeit und Wissen: Realismus, Antirealismus und Wirklichkeits-Konzeptionen in Philosophie und Wissenschaften. Hrsg.: H.J. Sandkühler
  18. Baumer, C.; Reichenberger, K.: Business Semantics - Praxis und Perspektiven (2006) 0.02
    0.018927898 = product of:
      0.07571159 = sum of:
        0.07571159 = weight(_text_:und in 20) [ClassicSimilarity], result of:
          0.07571159 = score(doc=20,freq=14.0), product of:
            0.14597435 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0658165 = queryNorm
            0.51866364 = fieldWeight in 20, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0625 = fieldNorm(doc=20)
      0.25 = coord(1/4)
    
    Abstract
    Der Artikel führt in semantische Technologien ein und gewährt Einblick in unterschiedliche Entwicklungsrichtungen. Insbesondere werden Business Semantics vorgestellt und vom Semantic Web abgegrenzt. Die Stärken von Business Semantics werden speziell an den Praxisbeispielen des Knowledge Portals und dem Projekt "Knowledge Base" der Wienerberger AG veranschaulicht. So werden die Anforderungen - was brauchen Anwendungen in Unternehmen heute - und die Leistungsfähigkeit von Systemen - was bieten Business Semantics - konkretisiert und gegenübergestellt.
    Source
    Information - Wissenschaft und Praxis. 57(2006) H.6/7, S.359-366
  19. Kunze, C.: Lexikalisch-semantische Wortnetze in Sprachwissenschaft und Sprachtechnologie (2006) 0.02
    0.018927898 = product of:
      0.07571159 = sum of:
        0.07571159 = weight(_text_:und in 23) [ClassicSimilarity], result of:
          0.07571159 = score(doc=23,freq=14.0), product of:
            0.14597435 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0658165 = queryNorm
            0.51866364 = fieldWeight in 23, product of:
              3.7416575 = tf(freq=14.0), with freq of:
                14.0 = termFreq=14.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0625 = fieldNorm(doc=23)
      0.25 = coord(1/4)
    
    Abstract
    Dieser Beitrag beschreibt die Strukturierungsprinzipien und Anwendungskontexte lexikalisch-semantischer Wortnetze, insbesondere des deutschen Wortnetzes GermaNet. Wortnetze sind zurzeit besonders populäre elektronische Lexikonressourcen, die große Abdeckungen semantisch strukturierter Datenfür verschiedene Sprachen und Sprachverbünde enthalten. In Wortnetzen sind die häufigsten und wichtigsten Konzepte einer Sprache mit ihren elementaren Bedeutungsrelationen repräsentiert. Zentrale Anwendungen für Wortnetze sind u.a. die Lesartendisambiguierung und die Informationserschließung. Der Artikel skizziert die neusten Szenarien, in denen GermaNet eingesetzt wird: die Semantische Informationserschließung und die Integration allgemeinsprachlicher Wortnetze mit terminologischen Ressourcen vordem Hintergrund der Datenkonvertierung in OWL.
    Source
    Information - Wissenschaft und Praxis. 57(2006) H.6/7, S.309-314
  20. Rahmstorf, G.: Strukturierung von inhaltlichen Daten : Topic Maps und Concepto (2004) 0.02
    0.018779442 = product of:
      0.07511777 = sum of:
        0.07511777 = weight(_text_:und in 4143) [ClassicSimilarity], result of:
          0.07511777 = score(doc=4143,freq=18.0), product of:
            0.14597435 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0658165 = queryNorm
            0.5145956 = fieldWeight in 4143, product of:
              4.2426405 = tf(freq=18.0), with freq of:
                18.0 = termFreq=18.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0546875 = fieldNorm(doc=4143)
      0.25 = coord(1/4)
    
    Abstract
    Topic Maps auf der einen Seite und das Programm Concepto auf der anderen Seite werden beschrieben. Mt Topic Maps können Wortnetze und einfache Satzstrukturen dargestellt werden. Concepto dient zur Erfassung, Bearbeitung und Visualisierung von Wortschatz und Strukturen. Es unterstützt ein Wortmodell, bei dem die verschiedenen Lesarten eines Wortes erfasst und einfachen, formalsprachlichen Begriffen zugewiesen werden können. Die Funktionen beider Systeme werden verglichen. Es wird dargestellt, was an Topic Maps und an Concepto ergänzt werden müsste, wenn beide Systeme einen kompatiblen, wechselseitigen Datenaustausch zulassen sollen. Diese Erweiterungen würden die Anwendbarkeit der Systeme noch interessanter machen.
    Source
    Wissensorganisation und Edutainment: Wissen im Spannungsfeld von Gesellschaft, Gestaltung und Industrie. Proceedings der 7. Tagung der Deutschen Sektion der Internationalen Gesellschaft für Wissensorganisation, Berlin, 21.-23.3.2001. Hrsg.: C. Lehner, H.P. Ohly u. G. Rahmstorf

Years

Languages

  • d 119
  • e 47
  • pt 1
  • More… Less…

Types

  • a 103
  • el 40
  • x 21
  • m 18
  • r 7
  • s 4
  • n 2
  • p 1
  • More… Less…

Subjects

Classifications