-
Cao, N.; Sun, J.; Lin, Y.-R.; Gotz, D.; Liu, S.; Qu, H.: FacetAtlas : Multifaceted visualization for rich text corpora (2010)
0.01
0.013681569 = product of:
0.054726277 = sum of:
0.054726277 = weight(_text_:have in 4366) [ClassicSimilarity], result of:
0.054726277 = score(doc=4366,freq=4.0), product of:
0.22215667 = queryWeight, product of:
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.07045517 = queryNorm
0.2463409 = fieldWeight in 4366, product of:
2.0 = tf(freq=4.0), with freq of:
4.0 = termFreq=4.0
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.0390625 = fieldNorm(doc=4366)
0.25 = coord(1/4)
- Abstract
- Documents in rich text corpora usually contain multiple facets of information. For example, an article about a specific disease often consists of different facets such as symptom, treatment, cause, diagnosis, prognosis, and prevention. Thus, documents may have different relations based on different facets. Powerful search tools have been developed to help users locate lists of individual documents that are most related to specific keywords. However, there is a lack of effective analysis tools that reveal the multifaceted relations of documents within or cross the document clusters. In this paper, we present FacetAtlas, a multifaceted visualization technique for visually analyzing rich text corpora. FacetAtlas combines search technology with advanced visual analytical tools to convey both global and local patterns simultaneously. We describe several unique aspects of FacetAtlas, including (1) node cliques and multifaceted edges, (2) an optimized density map, and (3) automated opacity pattern enhancement for highlighting visual patterns, (4) interactive context switch between facets. In addition, we demonstrate the power of FacetAtlas through a case study that targets patient education in the health care domain. Our evaluation shows the benefits of this work, especially in support of complex multifaceted data analysis.
-
Qu, R.; Fang, Y.; Bai, W.; Jiang, Y.: Computing semantic similarity based on novel models of semantic representation using Wikipedia (2018)
0.01
0.013681569 = product of:
0.054726277 = sum of:
0.054726277 = weight(_text_:have in 52) [ClassicSimilarity], result of:
0.054726277 = score(doc=52,freq=4.0), product of:
0.22215667 = queryWeight, product of:
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.07045517 = queryNorm
0.2463409 = fieldWeight in 52, product of:
2.0 = tf(freq=4.0), with freq of:
4.0 = termFreq=4.0
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.0390625 = fieldNorm(doc=52)
0.25 = coord(1/4)
- Abstract
- Computing Semantic Similarity (SS) between concepts is one of the most critical issues in many domains such as Natural Language Processing and Artificial Intelligence. Over the years, several SS measurement methods have been proposed by exploiting different knowledge resources. Wikipedia provides a large domain-independent encyclopedic repository and a semantic network for computing SS between concepts. Traditional feature-based measures rely on linear combinations of different properties with two main limitations, the insufficient information and the loss of semantic information. In this paper, we propose several hybrid SS measurement approaches by using the Information Content (IC) and features of concepts, which avoid the limitations introduced above. Considering integrating discrete properties into one component, we present two models of semantic representation, called CORM and CARM. Then, we compute SS based on these models and take the IC of categories as a supplement of SS measurement. The evaluation, based on several widely used benchmarks and a benchmark developed by ourselves, sustains the intuitions with respect to human judgments. In summary, our approaches are more efficient in determining SS between concepts and have a better human correlation than previous methods such as Word2Vec and NASARI.
-
Lobin, H.; Witt, A.: Semantic and thematic navigation in electronic encyclopedias (1999)
0.01
0.013544062 = product of:
0.05417625 = sum of:
0.05417625 = weight(_text_:have in 1624) [ClassicSimilarity], result of:
0.05417625 = score(doc=1624,freq=2.0), product of:
0.22215667 = queryWeight, product of:
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.07045517 = queryNorm
0.24386504 = fieldWeight in 1624, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.0546875 = fieldNorm(doc=1624)
0.25 = coord(1/4)
- Abstract
- In the field of electronic publishing, encyclopedias represent a unique sort of text for investigating advanced methods of navigation. The user of an electronic excyclopedia normally expects special methods for accessing the entries in an encyclopedia database. Navigation through printed encyclopedias in the traditional sense focuses on the alphabetic order of the entries. In electronic encyclopedias, however, thematic structuring of lemmas and, of course, extensive (hyper-) linking mechanisms have been added. This paper will focus on showing developments, which go beyond these navigational strucutres. We will concentrate on the semantic space formed by lemmas to build a network of semantic distances and thematic trails through the encyclopedia
-
Mlodzka-Stybel, A.: Towards continuous improvement of users' access to a library catalogue (2014)
0.01
0.013544062 = product of:
0.05417625 = sum of:
0.05417625 = weight(_text_:have in 2466) [ClassicSimilarity], result of:
0.05417625 = score(doc=2466,freq=2.0), product of:
0.22215667 = queryWeight, product of:
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.07045517 = queryNorm
0.24386504 = fieldWeight in 2466, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.0546875 = fieldNorm(doc=2466)
0.25 = coord(1/4)
- Abstract
- The paper discusses the issue of increasing users' access to library records by their publication in Google. Data from the records, converted into html format, have been indexed by Google. The process covered basic formal description fields of the records, description of the content, supported with a thesaurus, as well as an abstract, if present in the record. In addition to monitoring the end users' statistics, the pilot testing covered visibility of library records in Google search results.
-
Moreira, W.; Martínez-Ávila, D.: Concept relationships in knowledge organization systems : elements for analysis and common research among fields (2018)
0.01
0.013544062 = product of:
0.05417625 = sum of:
0.05417625 = weight(_text_:have in 166) [ClassicSimilarity], result of:
0.05417625 = score(doc=166,freq=2.0), product of:
0.22215667 = queryWeight, product of:
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.07045517 = queryNorm
0.24386504 = fieldWeight in 166, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.0546875 = fieldNorm(doc=166)
0.25 = coord(1/4)
- Abstract
- Knowledge organization systems have been studied in several fields and for different and complementary aspects. Among the aspects that concentrate common interests, in this article we highlight those related to the terminological and conceptual relationships among the components of any knowledge organization system. This research aims to contribute to the critical analysis of knowledge organization systems, especially ontologies, thesauri, and classification systems, by the comprehension of its similarities and differences when dealing with concepts and their ways of relating to each other as well as to the conceptual design that is adopted.
-
AssoziativOPAC : SpiderSearch von BOND (2003)
0.01
0.013538062 = product of:
0.054152247 = sum of:
0.054152247 = weight(_text_:und in 3029) [ClassicSimilarity], result of:
0.054152247 = score(doc=3029,freq=16.0), product of:
0.15626246 = queryWeight, product of:
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.07045517 = queryNorm
0.34654674 = fieldWeight in 3029, product of:
4.0 = tf(freq=16.0), with freq of:
16.0 = termFreq=16.0
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.0390625 = fieldNorm(doc=3029)
0.25 = coord(1/4)
- Content
- "Der Hersteller von Bibliothekssoftware BOND erweitert sein Produktangebot um den Assoziativ-OPAC SpiderSearch. Dieser graphische Web-OPAC sucht zu einem Stichwort Assoziative, d.h. sinnoder sprachwissenschaftlich verwandte Begriffe. Diese baut er spinnennetzartig um den zentralen Suchbegriff herum auf. Anhand der ihm angebotenen Assoziative kann sich der Leser sehr einfach und anschaulich durch den Medienbestand der Bibliothek klicken. So findet er schnell und komfortabel relevante Medien, die mit herkömmlichen Suchverfahren nur schwer recherchierbar wären. Überlegungen über verwandte Suchbegriffe und angrenzende Themengebiete bleiben dem Benutzer erspart: SpiderSearch navigiert den Benutzer ähnlich wie beim Surfen durch Webseiten durch sämtliche Themen, die mit dem Suchbegriff in Zusammenhang stehen. Auch aufwändiges Durchblättern einer riesigen Suchergebnisliste ist nicht nötig. Durch die im semantischen Netz vorgeschlagenen Begriffe kann der Benutzer sein Thema genau eingrenzen und erhält in seiner Trefferliste nur passende Medien. Diese ordnet SpiderSearch nach ihrer Relevanz, so dass der Leser die benötigte Literatur einfach und komfortabel findet. Wie auch im normalen Web-OPAC enthält die Trefferliste Angaben zu Titel, Standort und Verfügbarkeit des Mediums. Zur einfachen Zuordnung der Medienart ist jedem Medium ein entsprechendes Symbol zugewiesen. Per Mausklick erhält der Benutzer Detailangaben zum Medium und optional eine Ansicht des Buchcovers. SpiderSearch ist ein Zusatzmodul zur Software BIBLIOTHECA2000 von BOND und setzt auf den Web-OPAC auf."
-
Brunetti, J.M.; Roberto García, R.: User-centered design and evaluation of overview components for semantic data exploration (2014)
0.01
0.013405145 = product of:
0.05362058 = sum of:
0.05362058 = weight(_text_:have in 2626) [ClassicSimilarity], result of:
0.05362058 = score(doc=2626,freq=6.0), product of:
0.22215667 = queryWeight, product of:
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.07045517 = queryNorm
0.24136381 = fieldWeight in 2626, product of:
2.4494898 = tf(freq=6.0), with freq of:
6.0 = termFreq=6.0
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.03125 = fieldNorm(doc=2626)
0.25 = coord(1/4)
- Abstract
- Purpose - The growing volumes of semantic data available in the web result in the need for handling the information overload phenomenon. The potential of this amount of data is enormous but in most cases it is very difficult for users to visualize, explore and use this data, especially for lay-users without experience with Semantic Web technologies. The paper aims to discuss these issues. Design/methodology/approach - The Visual Information-Seeking Mantra "Overview first, zoom and filter, then details-on-demand" proposed by Shneiderman describes how data should be presented in different stages to achieve an effective exploration. The overview is the first user task when dealing with a data set. The objective is that the user is capable of getting an idea about the overall structure of the data set. Different information architecture (IA) components supporting the overview tasks have been developed, so they are automatically generated from semantic data, and evaluated with end-users. Findings - The chosen IA components are well known to web users, as they are present in most web pages: navigation bars, site maps and site indexes. The authors complement them with Treemaps, a visualization technique for displaying hierarchical data. These components have been developed following an iterative User-Centered Design methodology. Evaluations with end-users have shown that they get easily used to them despite the fact that they are generated automatically from structured data, without requiring knowledge about the underlying semantic technologies, and that the different overview components complement each other as they focus on different information search needs. Originality/value - Obtaining semantic data sets overviews cannot be easily done with the current semantic web browsers. Overviews become difficult to achieve with large heterogeneous data sets, which is typical in the Semantic Web, because traditional IA techniques do not easily scale to large data sets. There is little or no support to obtain overview information quickly and easily at the beginning of the exploration of a new data set. This can be a serious limitation when exploring a data set for the first time, especially for lay-users. The proposal is to reuse and adapt existing IA components to provide this overview to users and show that they can be generated automatically from the thesaurus and ontologies that structure semantic data while providing a comparable user experience to traditional web sites.
-
Weiermann, S.L.: Semantische Netze und Begriffsdeskription in der Wissensrepräsentation (2000)
0.01
0.013401995 = product of:
0.05360798 = sum of:
0.05360798 = weight(_text_:und in 4001) [ClassicSimilarity], result of:
0.05360798 = score(doc=4001,freq=8.0), product of:
0.15626246 = queryWeight, product of:
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.07045517 = queryNorm
0.34306374 = fieldWeight in 4001, product of:
2.828427 = tf(freq=8.0), with freq of:
8.0 = termFreq=8.0
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.0546875 = fieldNorm(doc=4001)
0.25 = coord(1/4)
- BK
- 18.00 Einzelne Sprachen und Literaturen allgemein
- Classification
- 18.00 Einzelne Sprachen und Literaturen allgemein
- Content
- Inhalt (in Kürze): Einleitung. Wissensrepräsentation. Semantische Netze. Wissensrepräsentationssysteme. Empirische Analyse und Systemerweiterungen.
-
Gödert, W.: Navigation und Konzepte für ein interaktives Retrieval im OPAC : Oder: Von der Informationserschließung zur Wissenserkundung (2004)
0.01
0.013264537 = product of:
0.053058147 = sum of:
0.053058147 = weight(_text_:und in 3195) [ClassicSimilarity], result of:
0.053058147 = score(doc=3195,freq=6.0), product of:
0.15626246 = queryWeight, product of:
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.07045517 = queryNorm
0.33954507 = fieldWeight in 3195, product of:
2.4494898 = tf(freq=6.0), with freq of:
6.0 = termFreq=6.0
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.0625 = fieldNorm(doc=3195)
0.25 = coord(1/4)
- Abstract
- In diesem Beitrag werden nach einem kurzen historischen Abriss der OPAC-Entwicklung die Prinzipien und Möglichkeiten zur Gestaltung von Interaktionsvorgängen zur inhaltlichen Suche diskutiert. Es wird dabei ein Plädoyer abgegeben, die OPACs nicht allein als Findeinstrumente für bibliografische Daten sondern auch als Systeme zur Wissenserkundung zu begreifen und die Interaktionsvorgänge darauf abzustimmen
-
Hoppe, T.: Semantische Filterung : ein Werkzeug zur Steigerung der Effizienz im Wissensmanagement (2013)
0.01
0.013264537 = product of:
0.053058147 = sum of:
0.053058147 = weight(_text_:und in 3245) [ClassicSimilarity], result of:
0.053058147 = score(doc=3245,freq=6.0), product of:
0.15626246 = queryWeight, product of:
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.07045517 = queryNorm
0.33954507 = fieldWeight in 3245, product of:
2.4494898 = tf(freq=6.0), with freq of:
6.0 = termFreq=6.0
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.0625 = fieldNorm(doc=3245)
0.25 = coord(1/4)
- Abstract
- Dieser Artikel adressiert einen Randbereich des Wissensmanagements: die Schnittstelle zwischen Unternehmens-externen Informationen im Internet und den Leistungsprozessen eines Unternehmens. Diese Schnittstelle ist besonders für Unternehmen von Interesse, deren Leistungsprozesse von externen Informationen abhängen und die auf diese Prozesse angewiesen sind. Wir zeigen an zwei Fallbeispielen, dass die inhaltliche Filterung von Informationen beim Eintritt ins Unternehmen ein wichtiges Werkzeug darstellt, um daran anschließende Wissens- und Informationsmanagementprozesse effizient zu gestalten.
-
Hauer, M: Silicon Valley Vorarlberg : Maschinelle Indexierung und semantisches Retrieval verbessert den Katalog der Vorarlberger Landesbibliothek (2004)
0.01
0.012663696 = product of:
0.050654784 = sum of:
0.050654784 = weight(_text_:und in 3489) [ClassicSimilarity], result of:
0.050654784 = score(doc=3489,freq=14.0), product of:
0.15626246 = queryWeight, product of:
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.07045517 = queryNorm
0.32416478 = fieldWeight in 3489, product of:
3.7416575 = tf(freq=14.0), with freq of:
14.0 = termFreq=14.0
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.0390625 = fieldNorm(doc=3489)
0.25 = coord(1/4)
- Abstract
- 10 Jahre Internet haben die WeIt um die Bibliotheken herum stark geändert. Der Web-OPAC war eine Antwort der Bibliotheken. Doch reicht ein Web-OPAC im Zeitalter des Internets noch aus? Außer Web ist es doch der alte Katalog. Ca. 90% aller Bibliotheksrecherchen durch Benutzer sind Themenrecherchen. Ein Anteil dieser Recherchen bringt kein Ergebnis. Es kann leicht gemessen werden, dass null Medien gefunden wurden. Die Gründe hierfür wurden auch immer wieder untersucht: Plural- anstelle Singularformen, zu spezifische Suchbegriffe, Schreib- oder Bedienungsfehler. Zu wenig untersucht sind aber die Recherchen, die nicht mit einer Ausleihe enden, denn auch dann kann man in vielen Fällen von einem Retrieval-Mangel ausgehen. Schließlich: Von den ausgeliehenen Büchern werden nach Einschätzung vieler Bibliothekare 80% nicht weiter als bis zum Inhaltsverzeichnis gelesen (außer in Präsenzbibliotheken) - und erst nach Wochen zurückgegeben. Ein Politiker würde dies neudeutsch als "ein Vermittlungsproblem" bezeichnen. Ein Controller als nicht hinreichende Kapitalnutzung. Einfacher machen es sich immer mehr Studenten und Wissenschaftler, ihr Wissensaustausch vollzieht sich zunehmend an anderen Orten. Bibliotheken (als Funktion) sind unverzichtbar für die wissenschaftliche Kommunikation. Deshalb geht es darum, Wege zu finden und auch zu beschreiten, welche die Schätze von Bibliotheken (als Institution) effizienter an die Zielgruppe bringen. Der Einsatz von Information Retrieval-Technologie, neue Erschließungsmethoden und neuer Content sind Ansätze dazu. Doch die bisherigen Verbundstrukturen und Abhängigkeit haben das hier vorgestellte innovative Projekt keineswegs gefördert. Innovation entsteht wie die Innvoationsforschung zeigt eigentlich immer an der Peripherie: in Bregenz fing es an.
- Source
- Mitteilungen der Vereinigung Österreichischer Bibliothekarinnen und Bibliothekare. 57(2004) H.3/4, S.33-38
-
Renker, L.: Exploration von Textkorpora : Topic Models als Grundlage der Interaktion (2015)
0.01
0.012663696 = product of:
0.050654784 = sum of:
0.050654784 = weight(_text_:und in 3380) [ClassicSimilarity], result of:
0.050654784 = score(doc=3380,freq=14.0), product of:
0.15626246 = queryWeight, product of:
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.07045517 = queryNorm
0.32416478 = fieldWeight in 3380, product of:
3.7416575 = tf(freq=14.0), with freq of:
14.0 = termFreq=14.0
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.0390625 = fieldNorm(doc=3380)
0.25 = coord(1/4)
- Abstract
- Das Internet birgt schier endlose Informationen. Ein zentrales Problem besteht heutzutage darin diese auch zugänglich zu machen. Es ist ein fundamentales Domänenwissen erforderlich, um in einer Volltextsuche die korrekten Suchanfragen zu formulieren. Das ist jedoch oftmals nicht vorhanden, so dass viel Zeit aufgewandt werden muss, um einen Überblick des behandelten Themas zu erhalten. In solchen Situationen findet sich ein Nutzer in einem explorativen Suchvorgang, in dem er sich schrittweise an ein Thema heranarbeiten muss. Für die Organisation von Daten werden mittlerweile ganz selbstverständlich Verfahren des Machine Learnings verwendet. In den meisten Fällen bleiben sie allerdings für den Anwender unsichtbar. Die interaktive Verwendung in explorativen Suchprozessen könnte die menschliche Urteilskraft enger mit der maschinellen Verarbeitung großer Datenmengen verbinden. Topic Models sind ebensolche Verfahren. Sie finden in einem Textkorpus verborgene Themen, die sich relativ gut von Menschen interpretieren lassen und sind daher vielversprechend für die Anwendung in explorativen Suchprozessen. Nutzer können damit beim Verstehen unbekannter Quellen unterstützt werden. Bei der Betrachtung entsprechender Forschungsarbeiten fiel auf, dass Topic Models vorwiegend zur Erzeugung statischer Visualisierungen verwendet werden. Das Sensemaking ist ein wesentlicher Bestandteil der explorativen Suche und wird dennoch nur in sehr geringem Umfang genutzt, um algorithmische Neuerungen zu begründen und in einen umfassenden Kontext zu setzen. Daraus leitet sich die Vermutung ab, dass die Verwendung von Modellen des Sensemakings und die nutzerzentrierte Konzeption von explorativen Suchen, neue Funktionen für die Interaktion mit Topic Models hervorbringen und einen Kontext für entsprechende Forschungsarbeiten bieten können.
- Footnote
- Masterthesis zur Erlangung des akademischen Grades Master of Science (M.Sc.) vorgelegt an der Fachhochschule Köln / Fakultät für Informatik und Ingenieurswissenschaften im Studiengang Medieninformatik.
- Imprint
- Gummersbach : Fakultät für Informatik und Ingenieurswissenschaften
-
Oard, D.W.: Alternative approaches for cross-language text retrieval (1997)
0.01
0.011729502 = product of:
0.04691801 = sum of:
0.04691801 = weight(_text_:have in 2164) [ClassicSimilarity], result of:
0.04691801 = score(doc=2164,freq=6.0), product of:
0.22215667 = queryWeight, product of:
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.07045517 = queryNorm
0.21119334 = fieldWeight in 2164, product of:
2.4494898 = tf(freq=6.0), with freq of:
6.0 = termFreq=6.0
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.02734375 = fieldNorm(doc=2164)
0.25 = coord(1/4)
- Abstract
- The explosive growth of the Internet and other sources of networked information have made automatic mediation of access to networked information sources an increasingly important problem. Much of this information is expressed as electronic text, and it is becoming practical to automatically convert some printed documents and recorded speech to electronic text as well. Thus, automated systems capable of detecting useful documents are finding widespread application. With even a small number of languages it can be inconvenient to issue the same query repeatedly in every language, so users who are able to read more than one language will likely prefer a multilingual text retrieval system over a collection of monolingual systems. And since reading ability in a language does not always imply fluent writing ability in that language, such users will likely find cross-language text retrieval particularly useful for languages in which they are less confident of their ability to express their information needs effectively. The use of such systems can be also be beneficial if the user is able to read only a single language. For example, when only a small portion of the document collection will ever be examined by the user, performing retrieval before translation can be significantly more economical than performing translation before retrieval. So when the application is sufficiently important to justify the time and effort required for translation, those costs can be minimized if an effective cross-language text retrieval system is available. Even when translation is not available, there are circumstances in which cross-language text retrieval could be useful to a monolingual user. For example, a researcher might find a paper published in an unfamiliar language useful if that paper contains references to works by the same author that are in the researcher's native language.
I will not attempt to draw a sharp distinction between retrieval and filtering in this survey. Although my own work on adaptive cross-language text filtering has led me to make this distinction fairly carefully in other presentations (c.f., (Oard 1997b)), such an proach does little to help understand the fundamental techniques which have been applied or the results that have been obtained in this case. Since it is still common to view filtering (detection of useful documents in dynamic document streams) as a kind of retrieval, will simply adopt that perspective here.
-
Rädler, K.: In Bibliothekskatalogen "googlen" : Integration von Inhaltsverzeichnissen, Volltexten und WEB-Ressourcen in Bibliothekskataloge (2004)
0.01
0.011724304 = product of:
0.046897218 = sum of:
0.046897218 = weight(_text_:und in 3432) [ClassicSimilarity], result of:
0.046897218 = score(doc=3432,freq=12.0), product of:
0.15626246 = queryWeight, product of:
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.07045517 = queryNorm
0.30011827 = fieldWeight in 3432, product of:
3.4641016 = tf(freq=12.0), with freq of:
12.0 = termFreq=12.0
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.0390625 = fieldNorm(doc=3432)
0.25 = coord(1/4)
- Abstract
- Ausgangslage Die Katalog-Recherchen über Internet, also von außerhalb der Bibliothek, nehmen erwartungsgemäß stark zu bzw. sind mittlerweile die Regel. Damit ist natürlich das Bedürfnis und die Notwendigkeit gewachsen, über den Titel hinaus zusätzliche inhaltliche Informationen zu erhalten, die es erlauben, die Zweckmäßigkeit wesentlich besser abschätzen zu können, eine Bestellung vorzunehmen oder vielleicht auch 50 km in die Bibliothek zu fahren, um ein Buch zu entleihen. Dieses Informationsdefizit wird zunehmend als gravierender Mangel erfahren. Inhaltsverzeichnisse referieren den Inhalt kurz und prägnant. Sie sind die erste Stelle, welche zur Relevanz-Beurteilung herangezogen wird. Fast alle relevanten Terme einer Fachbuchpublikation finden sich bereits dort. Andererseits wird immer deutlicher, dass die dem bibliothekarischen Paradigma entsprechende intellektuelle Indexierung der einzelnen dokumentarischen Einheiten mit den engsten umfassenden dokumentationssprachlichen Termen (Schlagwörter, Klassen) zwar eine notwendige, aber keinesfalls hinreichende Methode darstellt, das teuer erworbene Bibliotheksgut Information für den Benutzer in seiner spezifischen Problemstellung zu aktivieren und als Informationsdienstleistung anbieten zu können. Informationen zu sehr speziellen Fragestellungen, die oft nur in kürzeren Abschnitten (Kapitel) erörtert werden, sind derzeit nur indirekt, mit großem Zeitaufwand und oft überhaupt nicht auffindbar. Sie liegen sozusagen brach. Die Tiefe der intellektuellen Indexierung bis in einzelne inhaltliche Details zu erweitern, ist aus personellen und damit auch finanziellen Gesichtspunkten nicht vertretbar. Bibliotheken fallen deshalb in der Wahrnehmung von Informationssuchenden immer mehr zurück. Die enorme Informationsvielfalt liegt hinter dem Informations- bzw. Recherchehorizont der bibliographischen Aufnahmen im Katalog.
-
Fowler, R.H.; Wilson, B.A.; Fowler, W.A.L.: Information navigator : an information system using associative networks for display and retrieval (1992)
0.01
0.011609196 = product of:
0.046436783 = sum of:
0.046436783 = weight(_text_:have in 1919) [ClassicSimilarity], result of:
0.046436783 = score(doc=1919,freq=2.0), product of:
0.22215667 = queryWeight, product of:
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.07045517 = queryNorm
0.20902719 = fieldWeight in 1919, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.046875 = fieldNorm(doc=1919)
0.25 = coord(1/4)
- Abstract
- Document retrieval is a highly interactive process dealing with large amounts of information. Visual representations can provide both a means for managing the complexity of large information structures and an interface style well suited to interactive manipulation. The system we have designed utilizes visually displayed graphic structures and a direct manipulation interface style to supply an integrated environment for retrieval. A common visually displayed network structure is used for query, document content, and term relations. A query can be modified through direct manipulation of its visual form by incorporating terms from any other information structure the system displays. An associative thesaurus of terms and an inter-document network provide information about a document collection that can complement other retrieval aids. Visualization of these large data structures makes use of fisheye views and overview diagrams to help overcome some of the inherent difficulties of orientation and navigation in large information structures.
-
Nie, J.-Y.: Query expansion and query translation as logical inference (2003)
0.01
0.011609196 = product of:
0.046436783 = sum of:
0.046436783 = weight(_text_:have in 2425) [ClassicSimilarity], result of:
0.046436783 = score(doc=2425,freq=2.0), product of:
0.22215667 = queryWeight, product of:
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.07045517 = queryNorm
0.20902719 = fieldWeight in 2425, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.046875 = fieldNorm(doc=2425)
0.25 = coord(1/4)
- Abstract
- A number of studies have examined the problems of query expansion in monolingual Information Retrieval (IR), and query translation for crosslanguage IR. However, no link has been made between them. This article first shows that query translation is a special case of query expansion. There is also another set of studies an inferential IR. Again, there is no relationship established with query translation or query expansion. The second claim of this article is that logical inference is a general form that covers query expansion and query translation. This analysis provides a unified view of different subareas of IR. We further develop the inferential IR approach in two particular contexts: using fuzzy logic and probability theory. The evaluation formulas obtained are shown to strongly correspond to those used in other IR models. This indicates that inference is indeed the core of advanced IR.
-
Khan, M.S.; Khor, S.: Enhanced Web document retrieval using automatic query expansion (2004)
0.01
0.011609196 = product of:
0.046436783 = sum of:
0.046436783 = weight(_text_:have in 3091) [ClassicSimilarity], result of:
0.046436783 = score(doc=3091,freq=2.0), product of:
0.22215667 = queryWeight, product of:
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.07045517 = queryNorm
0.20902719 = fieldWeight in 3091, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.046875 = fieldNorm(doc=3091)
0.25 = coord(1/4)
- Abstract
- The ever growing popularity of the Internet as a source of information, coupled with the accompanying growth in the number of documents made available through the World Wide Web, is leading to an increasing demand for more efficient and accurate information retrieval tools. Numerous techniques have been proposed and tried for improving the effectiveness of searching the World Wide Web for documents relevant to a given topic of interest. The specification of appropriate keywords and phrases by the user is crucial for the successful execution of a query as measured by the relevance of documents retrieved. Lack of users' knowledge an the search topic and their changing information needs often make it difficult for them to find suitable keywords or phrases for a query. This results in searches that fail to cover all likely aspects of the topic of interest. We describe a scheme that attempts to remedy this situation by automatically expanding the user query through the analysis of initially retrieved documents. Experimental results to demonstrate the effectiveness of the query expansion scheure are presented.
-
Pahlevi, S.M.; Kitagawa, H.: Conveying taxonomy context for topic-focused Web search (2005)
0.01
0.011609196 = product of:
0.046436783 = sum of:
0.046436783 = weight(_text_:have in 4310) [ClassicSimilarity], result of:
0.046436783 = score(doc=4310,freq=2.0), product of:
0.22215667 = queryWeight, product of:
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.07045517 = queryNorm
0.20902719 = fieldWeight in 4310, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.046875 = fieldNorm(doc=4310)
0.25 = coord(1/4)
- Abstract
- Introducing context to a user query is effective to improve the search effectiveness. In this article we propose a method employing the taxonomy-based search services such as Web directories to facilitate searches in any Web search interfaces that support Boolean queries. The proposed method enables one to convey current search context an taxonomy of a taxonomy-based search service to the searches conducted with the Web search interfaces. The basic idea is to learn the search context in the form of a Boolean condition that is commonly accepted by many Web search interfaces, and to use the condition to modify the user query before forwarding it to the Web search interfaces. To guarantee that the modified query can always be processed by the Web search interfaces and to make the method adaptive to different user requirements an search result effectiveness, we have developed new fast classification learning algorithms.
-
Johnson, J.D.: On contexts of information seeking (2003)
0.01
0.011609196 = product of:
0.046436783 = sum of:
0.046436783 = weight(_text_:have in 2082) [ClassicSimilarity], result of:
0.046436783 = score(doc=2082,freq=2.0), product of:
0.22215667 = queryWeight, product of:
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.07045517 = queryNorm
0.20902719 = fieldWeight in 2082, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.046875 = fieldNorm(doc=2082)
0.25 = coord(1/4)
- Abstract
- While surprisingly little has been written about context at a meaningful level, context is central to most theoretical approaches to information seeking. In this essay I explore in more detail three senses of context. First, I look at context as equivalent to the situation in which a process is immersed. Second, I discuss contingency approaches that detail active ingredients of the situation that have specific, predictable effects. Third, I examine major frameworks for meaning systems. Then, I discuss how a deeper appreciation of context can enhance our understanding of the process of information seeking by examining two vastly different contexts in which it occurs: organizational and cancer-related, an exemplar of everyday life information seeking. This essay concludes with a discussion of the value that can be added to information seeking research and theory as a result of a deeper appreciation of context, particularly in terms of our current multi-contextual environment and individuals taking an active role in contextualizing.
-
Klas, C.-P.; Fuhr, N.; Schaefer, A.: Evaluating strategic support for information access in the DAFFODIL system (2004)
0.01
0.011609196 = product of:
0.046436783 = sum of:
0.046436783 = weight(_text_:have in 3419) [ClassicSimilarity], result of:
0.046436783 = score(doc=3419,freq=2.0), product of:
0.22215667 = queryWeight, product of:
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.07045517 = queryNorm
0.20902719 = fieldWeight in 3419, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.1531634 = idf(docFreq=5157, maxDocs=44421)
0.046875 = fieldNorm(doc=3419)
0.25 = coord(1/4)
- Abstract
- The digital library system Daffodil is targeted at strategic support of users during the information search process. For searching, exploring and managing digital library objects it provides user-customisable information seeking patterns over a federation of heterogeneous digital libraries. In this paper evaluation results with respect to retrieval effectiveness, efficiency and user satisfaction are presented. The analysis focuses on strategic support for the scientific work-flow. Daffodil supports the whole work-flow, from data source selection over information seeking to the representation, organisation and reuse of information. By embedding high level search functionality into the scientific work-flow, the user experiences better strategic system support due to a more systematic work process. These ideas have been implemented in Daffodil followed by a qualitative evaluation. The evaluation has been conducted with 28 participants, ranging from information seeking novices to experts. The results are promising, as they support the chosen model.