Search (1250 results, page 3 of 63)

  • × language_ss:"e"
  1. Zschunke, P.: Richtig googeln : Ein neues Buch hilft, alle Möglichkeiten der populären Suchmaschine zu nutzen (2003) 0.07
    0.06885915 = product of:
      0.1377183 = sum of:
        0.10366608 = weight(_text_:java in 55) [ClassicSimilarity], result of:
          0.10366608 = score(doc=55,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.23359407 = fieldWeight in 55, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0234375 = fieldNorm(doc=55)
        0.034052227 = weight(_text_:und in 55) [ClassicSimilarity], result of:
          0.034052227 = score(doc=55,freq=22.0), product of:
            0.13966292 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.06297082 = queryNorm
            0.24381724 = fieldWeight in 55, product of:
              4.690416 = tf(freq=22.0), with freq of:
                22.0 = termFreq=22.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.0234375 = fieldNorm(doc=55)
      0.5 = coord(2/4)
    
    Content
    "Fünf Jahre nach seiner Gründung ist Google zum Herz des weltweiten Computernetzes geworden. Mit seiner Konzentration aufs Wesentliche hat die Suchmaschine alle anderen Anbieter weit zurück gelassen. Aber Google kann viel mehr, als im Web nach Texten und Bildern zu suchen. Gesammelt und aufbereitet werden auch Beiträge in Diskussionsforen (Newsgroups), aktuelle Nachrichten und andere im Netz verfügbare Informationen. Wer sich beim "Googeln" darauf beschränkt, ein einziges Wort in das Suchformular einzutippen und dann die ersten von oft mehreren hunderttausend Treffern anzuschauen, nutzt nur einen winzigen Bruchteil der Möglichkeiten. Wie man Google bis zum letzten ausreizt, haben Tara Calishain und Rael Dornfest in einem bislang nur auf Englisch veröffentlichten Buch dargestellt (Tara Calishain/Rael Dornfest: Google Hacks", www.oreilly.de, 28 Euro. Die wichtigsten Praxistipps kosten als Google Pocket Guide 12 Euro). - Suchen mit bis zu zehn Wörtern - Ihre "100 Google Hacks" beginnen mit Google-Strategien wie der Kombination mehrerer Suchbegriffe und enden mit der Aufforderung zur eigenen Nutzung der Google API ("Application Programming Interface"). Diese Schnittstelle kann zur Entwicklung von eigenen Programmen eingesetzt werden,,die auf die Google-Datenbank mit ihren mehr als drei Milliarden Einträgen zugreifen. Ein bewussteres Suchen im Internet beginnt mit der Kombination mehrerer Suchbegriffe - bis zu zehn Wörter können in das Formularfeld eingetippt werden, welche Google mit dem lo-gischen Ausdruck "und" verknüpft. Diese Standardvorgabe kann mit einem dazwischen eingefügten "or" zu einer Oder-Verknüpfung geändert werden. Soll ein bestimmter Begriff nicht auftauchen, wird ein Minuszeichen davor gesetzt. Auf diese Weise können bei einer Suche etwa alle Treffer ausgefiltert werden, die vom Online-Buchhändler Amazon kommen. Weiter gehende Syntax-Anweisungen helfen ebenfalls dabei, die Suche gezielt einzugrenzen: Die vorangestellte Anweisung "intitle:" etwa (ohne Anführungszeichen einzugeben) beschränkt die Suche auf all diejenigen Web-Seiten, die den direkt danach folgenden Begriff in ihrem Titel aufführen. Die Computer von Google bewältigen täglich mehr als 200 Millionen Anfragen. Die Antworten kommen aus einer Datenbank, die mehr als drei Milliarden Einträge enthält und regelmäßig aktualisiert wird. Dazu Werden SoftwareRoboter eingesetzt, so genannte "Search-Bots", die sich die Hyperlinks auf Web-Seiten entlang hangeln und für jedes Web-Dokument einen Index zur Volltextsuche anlegen. Die Einnahmen des 1998 von Larry Page und Sergey Brin gegründeten Unternehmens stammen zumeist von Internet-Portalen, welche die GoogleSuchtechnik für ihre eigenen Dienste übernehmen. Eine zwei Einnahmequelle ist die Werbung von Unternehmen, die für eine optisch hervorgehobene Platzierung in den GoogleTrefferlisten zahlen. Das Unternehmen mit Sitz im kalifornischen Mountain View beschäftigt rund 800 Mitarbeiter. Der Name Google leitet sich ab von dem Kunstwort "Googol", mit dem der amerikanische Mathematiker Edward Kasner die unvorstellbar große Zahl 10 hoch 100 (eine 1 mit hundert Nullen) bezeichnet hat. Kommerzielle Internet-Anbieter sind sehr, daran interessiert, auf den vordersten Plätzen einer Google-Trefferliste zu erscheinen.
    Da Google im Unterschied zu Yahoo oder Lycos nie ein auf möglichst viele Besuche angelegtes Internet-Portal werden wollte, ist die Suche in der Datenbank auch außerhalb der Google-Web-Site möglich. Dafür gibt es zunächst die "Google Toolbar" für den Internet Explorer, mit der dieser Browser eine eigene Leiste, für die Google-Suche erhält. Freie Entwickler bieten im Internet eine eigene Umsetzung: dieses Werkzeugs auch für den Netscape/ Mozilla-Browser an. Daneben kann ein GoogleSucheingabefeld aber auch auf die eigene WebSeite platziert werden - dazu sind nur vier Zei-len HTML-Code nötig. Eine Google-Suche zu starten, ist übrigens auch ganz ohne Browser möglich. Dazu hat das Unternehmen im Aprilvergangenen Jahres die API ("Application Programming Interface") frei gegeben, die in eigene Programme' eingebaut wird. So kann man etwa eine Google-Suche mit einer E-Mail starten: Die Suchbegriffe werden in die Betreff Zeile einer ansonsten leeren EMail eingetragen, die an die Adresse google@capeclear.com geschickt wird. Kurz danach trifft eine automatische Antwort-Mail mit den ersten zehn Treffern ein. Die entsprechenden Kenntnisse vorausgesetzt, können Google-Abfragen auch in Web-Services eingebaut werden - das sind Programme, die Daten aus dem Internet verarbeiten. Als Programmiertechniken kommen dafür Perl, PHP, Python oder Java in Frage. Calishain und Dornfest stellen sogar eine Reihe von abgedrehten Sites vor, die solche Programme für abstrakte Gedichte oder andere Kunstwerke einsetzen."
  2. Gillman, P.: Data handling and text compression (1992) 0.06
    0.06332112 = product of:
      0.25328448 = sum of:
        0.25328448 = weight(_text_:handled in 5305) [ClassicSimilarity], result of:
          0.25328448 = score(doc=5305,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.5163721 = fieldWeight in 5305, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.046875 = fieldNorm(doc=5305)
      0.25 = coord(1/4)
    
    Abstract
    Data compression has a function in text storage and data handling, but not at the level of compressing data files. The reason is that the decompression of such files add a time delay to the retrieval process, and users can see this delay as a drawback of the system concerned. Compression techniques can with benefit be applied to index files. A more relevant data handling problem is that posed by the need, in most systems, to store two versions of imported text. The first id the 'native' version, as it might have come from a word processor or text editor. The second is the ASCII version which is what is actually imported. Inverted file indexes form yet another version. The problem arises out of the need for dynamic indexing and re-indexing of revisable documents in very large database applications such as are found in Office Automation systems. Four mainstream text-management packages are used to show how this problem is handled, and how generic document architectures such as OCA/CDA and SGML might help
  3. Shemberg, M.: From citation to piece in hand : the search for efficiency in accessing monographic series (1997) 0.06
    0.06332112 = product of:
      0.25328448 = sum of:
        0.25328448 = weight(_text_:handled in 1539) [ClassicSimilarity], result of:
          0.25328448 = score(doc=1539,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.5163721 = fieldWeight in 1539, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.046875 = fieldNorm(doc=1539)
      0.25 = coord(1/4)
    
    Abstract
    Looks at access problems that occur when the citations from indexes for articles in series differ from the way the items are catalogued. To illustrate the difficulties encountered,a search was conducted in the online catalogues of various libraries and a comparison made between the bibliographiuc entries found to citations both in electronic indexes and in authors' references. The series: New directions for community colleges (NDCC); was chosen as it has kept the same title, ISSN and editors since 1973; yet is still illusive to identify. Each issue of NDCC, beginning with issue 1, 1973, was examined for citation information, numbering systems and changes in description on the title page. The online catalogue at Ohio State University (OSCAR) was checked to see how each issue was handled over time. Indexes consulted included: ERIC, Education Index; ArticleFirst; and CARL UnCover. Results indicate that the ramifications of local cataloguing decisions affect all areas of librarianship. In this electronic age, the need is great to integrate the various sources of information needed for a patron to go from citation to piece in hand
  4. Buizza, P.; Guerrini, M: ¬A conceptual model for the new Soggettario : subject indexing in the light of FRBR (2002) 0.06
    0.06332112 = product of:
      0.25328448 = sum of:
        0.25328448 = weight(_text_:handled in 685) [ClassicSimilarity], result of:
          0.25328448 = score(doc=685,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.5163721 = fieldWeight in 685, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.046875 = fieldNorm(doc=685)
      0.25 = coord(1/4)
    
    Content
    The National Central Library in Florence, Italy, has commissioned a feasibility study for the renewal of the Soggettario [Subject headings for Italian libraries]. [It is indispensable for the theoretical development to take place within the international debate and to approach the topic of a new Soggettario with reference to the FRBR.}. The subject is analysed as a relation between the entities in the third group: concept, object, event, place and the entity work. The model identifies the logical entities, attributes and relationships which run between the entities. The article returns to and amplifies the user tasks of FRBR which involve a subject: (1) Find the works on a given subject; (2) Find the works in which a concept is significantly treated; (3) Select a work by its main subject only; (4) Lead to a search for works on related subjects; (5) Lead to a search for works in which related or connected subjects are handled.
  5. Souza, R.R.; Raghavan, K.S.: ¬A methodology for noun phrase-based automatic indexing (2006) 0.06
    0.06332112 = product of:
      0.25328448 = sum of:
        0.25328448 = weight(_text_:handled in 298) [ClassicSimilarity], result of:
          0.25328448 = score(doc=298,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.5163721 = fieldWeight in 298, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.046875 = fieldNorm(doc=298)
      0.25 = coord(1/4)
    
    Abstract
    The scholarly community is increasingly employing the Web both for publication of scholarly output and for locating and accessing relevant scholarly literature. Organization of this vast body of digital information assumes significance in this context. The sheer volume of digital information to be handled makes traditional indexing and knowledge representation strategies ineffective and impractical. It is, therefore, worth exploring new approaches. An approach being discussed considers the intrinsic semantics of texts of documents. Based on the hypothesis that noun phrases in a text are semantically rich in terms of their ability to represent the subject content of the document, this approach seeks to identify and extract noun phrases instead of single keywords, and use them as descriptors. This paper presents a methodology that has been developed for extracting noun phrases from Portuguese texts. The results of an experiment carried out to test the adequacy of the methodology are also presented.
  6. Rodriguez Bravo, B.: ¬The visibility of women in indexing languages (2006) 0.06
    0.06332112 = product of:
      0.25328448 = sum of:
        0.25328448 = weight(_text_:handled in 1263) [ClassicSimilarity], result of:
          0.25328448 = score(doc=1263,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.5163721 = fieldWeight in 1263, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.046875 = fieldNorm(doc=1263)
      0.25 = coord(1/4)
    
    Abstract
    This article analyses how gender matters are handled in indexing languages. The examples chosen were the Library of Congress Subject Headings (LCSH), the UNESCO Thesaurus (UT) and the European Women's Thesaurus (EWT). The study is based on an analysis of the entries Man/Men and Woman/Women, their subdivisions and established relationship appearing under these entries. Other headings or descriptors are also listed when they allude to men or women but the gender sense occupies only second or third place in the entry, in the shape of an adjective or a second noun. A lack of symmetry, in the treatment of gender is noted, with recommendations being made for equal status for men and women, which should, however, avoid unnecessary enumerations.
  7. Salembier, P.; Benitez, A.B.: Structure description tools (2007) 0.06
    0.06332112 = product of:
      0.25328448 = sum of:
        0.25328448 = weight(_text_:handled in 1474) [ClassicSimilarity], result of:
          0.25328448 = score(doc=1474,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.5163721 = fieldWeight in 1474, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.046875 = fieldNorm(doc=1474)
      0.25 = coord(1/4)
    
    Abstract
    This article provides an overview of the tools specified by the MPEG-7 standard for describing the structure of multimedia content. In particular, it focuses on tools that represent segments resulting from a spatial and/or temporal partitioning of multimedia content. The segments are described in terms of their decomposition and the general relations among them as well as attributes or features of segments. Decomposition efficiently represents segment hierarchies and can be used to create tables of contents or indexes. More general graph representations are handled by the various standard spatial and temporal relations. A segment can be described by a large number of features ranging from those targeting the life cycle of the content (e.g., creation and usage) to those addressing signal characteristics such as audio, color, shape, or motion properties.
  8. Broughton, V.: ¬A faceted classification as the basis of a faceted terminology : conversion of a classified structure to thesaurus format in the Bliss Bibliographic Classification, 2nd Edition (2008) 0.06
    0.06332112 = product of:
      0.25328448 = sum of:
        0.25328448 = weight(_text_:handled in 2857) [ClassicSimilarity], result of:
          0.25328448 = score(doc=2857,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.5163721 = fieldWeight in 2857, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.046875 = fieldNorm(doc=2857)
      0.25 = coord(1/4)
    
    Abstract
    Facet analysis is an established methodology for building classifications and subject indexing systems, but has been less rigorously applied to thesauri. The process of creating a compatible thesaurus from the schedules of the Bliss Bibliographic Classification 2nd edition highlights the ways in which the conceptual relationships in a subject field are handled in the two types of retrieval languages. An underlying uniformity of theory is established, and the way in which software can manage the relationships is discussed. The manner of displaying verbal expressions of concepts (vocabulary control) is also considered, but is found to be less well controlled in the classification than in the thesaurus. Nevertheless, there is good reason to think that facet analysis provides a sound basis for structuring a variety of knowledge organization tools.
  9. Han, Y.: ¬A RDF-based digital library system (2006) 0.06
    0.06332112 = product of:
      0.25328448 = sum of:
        0.25328448 = weight(_text_:handled in 3524) [ClassicSimilarity], result of:
          0.25328448 = score(doc=3524,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.5163721 = fieldWeight in 3524, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.046875 = fieldNorm(doc=3524)
      0.25 = coord(1/4)
    
    Abstract
    Purpose - To research a resource description framework (RDF) based digital library system that facilitates digital resource management and supports knowledge management for an interoperable information environment. Design/methodology/approach - The paper first introduces some of issues with metadata management and knowledge management and describes the needs for a true interoperable environment for information transferring across domains. A journal delivery application has been implemented as a concept-proof project to demonstrate the usefulness of RDF in digital library systems. Findings - The RDF-based digital library system at the University of Arizona Libraries provides an easy way for digital resource management by integrating other applications regardless of metadata formats and web presence. Practical implications - A journal delivery application has been running in the RDF-based digital library system since April 2005. An electronic theses and dissertation application will be handled by the same system. Originality/value - The paper suggests to use RDF, the semantic web technology, as a new approach to facilitate knowledge management and metadata management. Using RDF technology brings new ways to manage and discover information for libraries.
  10. Genereux, C.: Building connections : a review of the serials literature 2004 through 2005 (2007) 0.06
    0.06332112 = product of:
      0.25328448 = sum of:
        0.25328448 = weight(_text_:handled in 3548) [ClassicSimilarity], result of:
          0.25328448 = score(doc=3548,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.5163721 = fieldWeight in 3548, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.046875 = fieldNorm(doc=3548)
      0.25 = coord(1/4)
    
    Abstract
    This review of 2004 and 2005 serials literature covers the themes of cost, management, and access. Interwoven through the serials literature of these two years are the importance of collaboration, communication, and linkages between scholars, publishers, subscription agents and other intermediaries, and librarians. The emphasis in the literature is on electronic serials and their impact on publishing, libraries, and vendors. In response to the crisis of escalating journal prices and libraries' dissatisfaction with the Big Deal licensing agreements, Open Access journals and publishing models were promoted. Libraries subscribed to or licensed increasing numbers of electronic serials. As a result, libraries sought ways to better manage licensing and subscription data (not handled by traditional integrated library systems) by implementing electronic resources management systems. In order to provide users with better, faster, and more current information on and access to electronic serials, libraries implemented tools and services to provide A-Z title lists, title by title coverage data, MARC records, and OpenURL link resolvers.
  11. Lee, Y.Y.; Yang, S.Q.: Folksonomies as subject access : a survey of tagging in library online catalogs and discovery layers (2012) 0.06
    0.06332112 = product of:
      0.25328448 = sum of:
        0.25328448 = weight(_text_:handled in 1309) [ClassicSimilarity], result of:
          0.25328448 = score(doc=1309,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.5163721 = fieldWeight in 1309, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.046875 = fieldNorm(doc=1309)
      0.25 = coord(1/4)
    
    Abstract
    This paper describes a survey on how system vendors and libraries handled tagging in OPACs and discovery layers. Tags are user added subject metadata, also called folksonomies. This survey also investigated user behavior when they face the possibility to tag. The findings indicate that legacy/classic systems have no tagging capability. About 47% of the discovery tools provide tagging function. About 49% of the libraries that have a system with tagging capability have turned the tagging function on in their OPACs and discovery tools. Only 40% of the libraries that turned tagging on actually utilized user added subject metadata as access point to collections. Academic library users are less active in tagging than public library users.
  12. Chu, H.: Information representation and retrieval in the digital age (2010) 0.06
    0.06303159 = product of:
      0.12606318 = sum of:
        0.04163502 = weight(_text_:und in 1092) [ClassicSimilarity], result of:
          0.04163502 = score(doc=1092,freq=74.0), product of:
            0.13966292 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.06297082 = queryNorm
            0.29811078 = fieldWeight in 1092, product of:
              8.602325 = tf(freq=74.0), with freq of:
                74.0 = termFreq=74.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.015625 = fieldNorm(doc=1092)
        0.08442816 = weight(_text_:handled in 1092) [ClassicSimilarity], result of:
          0.08442816 = score(doc=1092,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.17212403 = fieldWeight in 1092, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.015625 = fieldNorm(doc=1092)
      0.5 = coord(2/4)
    
    Footnote
    Rez. in: JASIST 56(2005) no.2, S.215-216 (A. Heath): "What is small, thoroughly organized, and easy to understand? Well, it's Heting Chu's latest book an information retrieval. A very welcome release, this small literary addition to the field (only 248 pages) contains a concise and weIl-organized discussion of every major topic in information retrieval. The often-complex field of information retrieval is presented from its origin in the early 1950s to the present day. The organization of this text is top-notch, thus making this an easy read for even the novice. Unlike other titles in this area, Chu's user-friendly style of writing is done an purpose to properly introduce newcomers to the field in a less intimidating way. As stated by the author in the Preface, the purpose of the book is to "present a systematic, thorough yet nontechnical view of the field by using plain language to explain complex subjects." Chu has definitely struck up the right combination of ingredients. In a field so broad and complex, a well-organized presentation of topics that don't trip an themselves is essential. The use of plain language where possible is also a good choice for this topic because it allows one to absorb topics that are, by nature, not as easy to grasp. For instance, Chapters 6 and 7, which cover retrieval approaches and techniques, an often painstaking topic for many students and teachers is deftly handled with the use of tables that can be used to compare and contrast the various models discussed. I particularly loved Chu's use of Koll's 2000 article from the Bulletin of the American Society for Information Science to explain subject searching at the beginning of Chapter 6, which discusses the differences between browsing and searching. The Koll article uses the task of finding a needle in a haystack as an analogy.
    Weitere Rez. in: Rez. in: nfd 55(2004) H.4, S.252 (D. Lewandowski):"Die Zahl der Bücher zum Thema Information Retrieval ist nicht gering, auch in deutscher Sprache liegen einige Titel vor. Trotzdem soll ein neues (englischsprachiges) Buch zu diesem Thema hier besprochen werden. Dieses zeichnet sich durch eine Kürze (nur etwa 230 Seiten Text) und seine gute Verständlichkeit aus und richtet sich damit bevorzugt an Studenten in den ersten Semestern. Heting Chu unterrichtet seit 1994 an Palmer School of Library and Information Science der Long Island University New York. Dass die Autorin viel Erfahrung in der Vermittlung des Stoffs in ihren Information-Retrieval-Veranstaltungen sammeln konnte, merkt man dem Buch deutlich an. Es ist einer klaren und verständlichen Sprache geschrieben und führt in die Grundlagen der Wissensrepräsentation und des Information Retrieval ein. Das Lehrbuch behandelt diese Themen als Gesamtkomplex und geht damit über den Themenbereich ähnlicher Bücher hinaus, die sich in der Regel auf das Retrieval beschränken. Das Buch ist in zwölf Kapitel gegliedert, wobei das erste Kapitel eine Übersicht über die zu behandelnden Themen gibt und den Leser auf einfache Weise in die Grundbegriffe und die Geschichte des IRR einführt. Neben einer kurzen chronologischen Darstellung der Entwicklung der IRR-Systeme werden auch vier Pioniere des Gebiets gewürdigt: Mortimer Taube, Hans Peter Luhn, Calvin N. Mooers und Gerard Salton. Dies verleiht dem von Studenten doch manchmal als trocken empfundenen Stoff eine menschliche Dimension. Das zweite und dritte Kapitel widmen sich der Wissensrepräsentation, wobei zuerst die grundlegenden Ansätze wie Indexierung, Klassifikation und Abstracting besprochen werden. Darauf folgt die Behandlung von Wissensrepräsentation mittels Metadaten, wobei v.a. neuere Ansätze wie Dublin Core und RDF behandelt werden. Weitere Unterkapitel widmen sich der Repräsentation von Volltexten und von Multimedia-Informationen. Die Stellung der Sprache im IRR wird in einem eigenen Kapitel behandelt. Dabei werden in knapper Form verschiedene Formen des kontrollierten Vokabulars und die wesentlichen Unterscheidungsmerkmale zur natürlichen Sprache erläutert. Die Eignung der beiden Repräsentationsmöglichkeiten für unterschiedliche IRR-Zwecke wird unter verschiedenen Aspekten diskutiert.
    Die Kapitel fünf bis neun widmen sich dann ausführlich dem Information Retrieval. Zuerst werden grundlegende Retrievaltechniken vorgestellt und ihre Vor- und Nachteile dargestellt. Aus der Sicht des Nutzers von IR-Systemen wird der Prozess des Bildens einer Suchanfrage diskutiert und die damit verbundenen Probleme aufgezeigt. Im sechsten Kapitel werden die Retrieval-Ansätze Suchen und Browsen gegenübergestellt, entsprechende Suchstrategien aufgezeigt und schließlich Ansätze diskutiert, die suchen und browsen zu integrieren versuchen. Das siebte Kapitel beschäftigt sich dann mit dem, was den Kern der meisten IRBücher ausmacht: den IR-Modellen. Diese werden kurz vorgestellt, auf Formeln wird weitgehend verzichtet. Dies ist jedoch durchaus als Vorteil zu sehen, denn gerade Studienanfängern bereitet das Verständnis der IR-Modelle aufgrund deren Komplexität oft Schwierigkeiten. Nach der Lektüre dieses Kapitels wird man zwar nicht im Detail über die verschiedenen Modelle Bescheid wissen, wird sie jedoch kennen und einordnen können.
    In Kapitel acht werden unterschiedliche Arten von IR-Systemen vorgestellt. Dies sind Online IR-Systeme, CD-ROM-Systeme, OPACs und Internet IR-Systeme, denen der Grossteil dieses Kapitels gewidmet ist. Zu jeder Art von System werden die historische Entwicklung und die Besonderheiten genannt. Bei den Internet-IR-Systemen wird ausführlich auf die besonderen Probleme, die bei diesen im Vergleich zu klassischen IR-Systemen auftauchen, eingegangen. Ein extra Kapitel behandelt die Besonderheiten des Retrievals bei besonderen Dokumentkollektionen und besonderen Formaten. Hier finden sich Informationen zum multilingualen Retrieval und zur Suche nach Multimedia-Inhalten, wobei besonders auf die Unterscheidung zwischen beschreibungs- und inhaltsbasiertem Ansatz der Erschließung solcher Inhalte eingegangen wird. In Kapitel zehn erfährt der Leser mehr über die Stellung des Nutzers in IRR-Prozessen. Die Autorin stellt verschiedene Arten von Suchinterfaces bzw. Benutzeroberflächen und Ansätze der Evaluation der Mensch-Maschine-Interaktion in solchen Systemen vor. Kapitel elf beschäftigt sich ausführlich mit der Evaluierung von IRR-Systemen und stellt die bedeutendsten Test (Cranfield und TREC) vor Ein kurzes abschließendes Kapitel behandelt Ansätze der künstlichen Intelligenz und ihre Anwendung bei IRR-Systemen. Der Aufbau, die knappe, aber dennoch präzise Behandlung des Themas sowie die verständliche Sprache machen dieses Buch zu eine sehr guten Einführung für Studenten in den ersten Semestern, die der englischen Sprache mächtig sind. Besonders positiv hervorzuheben ist die Behandlung auch der aktuellen Themen des IRR wie der Einsatz von Metadaten, die Behandlung von Multimedia-Informationen und der Schwerpunk bei den Internet-IR-Systemen.
    Leider gibt es in deutscher Sprache keinen vergleichbaren Titel. Das Information-Retrieval-Buch von Ferber (2003) ist eher mathematisch orientiert und dürfte Studienanfänger der Informationswissenschaft durch seine große Detailliertheit und der damit einhergehenden großen Anzahl von Formeln eher abschrecken. Es ist eher denjenigen empfohlen, die sich intensiver mit dem Thema beschäftigen möchten. Ähnlich verhält es sich mit dem von manchen gerne genutzten Skript von Fuhr. Das Buch von Gaus (2003) ist mittlerweile schon ein Klassiker, beschäftigt sich aber im wesentlichen mit der Wissensrepräsentation und bietet zudem wenig Aktuelles. So fehlen etwa die Themen Information Retrieval im Internet und Multimedia-Retrieval komplett. Auch die Materialsammlung von Poetzsch (2002) konzentriert sich auf IR in klassischen Datenbanken und strebt zudem auch keine systematische Darstellung des Gebiets an. Zu wünschen wäre also, dass das hier besprochene Buch auch hierzulande in der Lehre Verwendung finden würde, da es den Studierenden einen knappen, gut lesbaren Einblick in das Themengebiet gibt. Es sollte aufgrund der vorbildlichen Aufbereitung des Stoffs auch Vorbild für zukünftige Autoren von Lehrbüchern sein. Und letztlich würde sich der Rezensent eine deutsche Übersetzung dieses Bandes wünschen."
  13. Braeckman, J.: ¬The integration of library information into a campus wide information system (1996) 0.06
    0.06047188 = product of:
      0.24188752 = sum of:
        0.24188752 = weight(_text_:java in 729) [ClassicSimilarity], result of:
          0.24188752 = score(doc=729,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.5450528 = fieldWeight in 729, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0546875 = fieldNorm(doc=729)
      0.25 = coord(1/4)
    
    Abstract
    Discusses the development of Campus Wide Information Systems with reference to the work of Leuven University Library. A 4th phase can now be distinguished in the evolution of CWISs as they evolve towards Intranets. WWW technology is applied to organise a consistent interface to different types of information, databases and services within an institution. WWW servers now exist via which queries and query results are translated from the Web environment to the specific database query language and vice versa. The integration of Java will enable programs to be executed from within the Web environment. Describes each phase of CWIS development at KU Leuven
  14. Chang, S.-F.; Smith, J.R.; Meng, J.: Efficient techniques for feature-based image / video access and manipulations (1997) 0.06
    0.06047188 = product of:
      0.24188752 = sum of:
        0.24188752 = weight(_text_:java in 756) [ClassicSimilarity], result of:
          0.24188752 = score(doc=756,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.5450528 = fieldWeight in 756, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0546875 = fieldNorm(doc=756)
      0.25 = coord(1/4)
    
    Abstract
    Describes 2 research projects aimed at studying the parallel issues of image and video indexing, information retrieval and manipulation: VisualSEEK, a content based image query system and a Java based WWW application supporting localised colour and spatial similarity retrieval; and CVEPS (Compressed Video Editing and Parsing System) which supports video manipulation with indexing support of individual frames from VisualSEEK and a hierarchical new video browsing and indexing system. In both media forms, these systems address the problem of heterogeneous unconstrained collections
  15. Lo, M.L.: Recent strategies for retrieving chemical structure information on the Web (1997) 0.06
    0.06047188 = product of:
      0.24188752 = sum of:
        0.24188752 = weight(_text_:java in 3611) [ClassicSimilarity], result of:
          0.24188752 = score(doc=3611,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.5450528 = fieldWeight in 3611, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0546875 = fieldNorm(doc=3611)
      0.25 = coord(1/4)
    
    Abstract
    Discusses various structural searching methods available on the Web. some databases such as the Brookhaven Protein Database use keyword searching which does not provide the desired substructure search capabilities. Others like CS ChemFinder and MDL's Chemscape use graphical plug in programs. Although plug in programs provide more capabilities, users first have to obtain a copy of the programs. Due to this limitation, Tripo's WebSketch and ACD Interactive Lab adopt a different approach. Using JAVA applets, users create and display a structure query of the molecule on the web page without using other software. The new technique is likely to extend itself to other electronic publications
  16. Kirschenbaum, M.: Documenting digital images : textual meta-data at the Blake Archive (1998) 0.06
    0.06047188 = product of:
      0.24188752 = sum of:
        0.24188752 = weight(_text_:java in 4287) [ClassicSimilarity], result of:
          0.24188752 = score(doc=4287,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.5450528 = fieldWeight in 4287, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0546875 = fieldNorm(doc=4287)
      0.25 = coord(1/4)
    
    Abstract
    Describes the work undertaken by the Wiliam Blake Archive, Virginia University, to document the metadata tools for handling digital images of illustrations accompanying Blake's work. Images are encoded in both JPEG and TIFF formats. Image Documentation (ID) records are slotted into that portion of the JPEG file reserved for textual metadata. Because the textual content of the ID record now becomes part of the image file itself, the documentary metadata travels with the image even it it is downloaded from one file to another. The metadata is invisible when viewing the image but becomes accessible to users via the 'info' button on the control panel of the Java applet
  17. Priss, U.: ¬A graphical interface for conceptually navigating faceted thesauri (1998) 0.06
    0.06047188 = product of:
      0.24188752 = sum of:
        0.24188752 = weight(_text_:java in 658) [ClassicSimilarity], result of:
          0.24188752 = score(doc=658,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.5450528 = fieldWeight in 658, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0546875 = fieldNorm(doc=658)
      0.25 = coord(1/4)
    
    Abstract
    This paper describes a graphical interface for the navigation and construction of faceted thesauri that is based on formal concept analysis. Each facet of a thesaurus is represented as a mathematical lattice that is further subdivided into components. Users can graphically navigate through the Java implementation of the interface by clicking on terms that connect facets and components. Since there are many applications for thesauri in the knowledge representation field, such a graphical interface has the potential of being very useful
  18. Renehan, E.J.: Science on the Web : a connoisseur's guide to over 500 of the best, most useful, and most fun science Websites (1996) 0.06
    0.06047188 = product of:
      0.24188752 = sum of:
        0.24188752 = weight(_text_:java in 1211) [ClassicSimilarity], result of:
          0.24188752 = score(doc=1211,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.5450528 = fieldWeight in 1211, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0546875 = fieldNorm(doc=1211)
      0.25 = coord(1/4)
    
    Abstract
    Written by the author of the best-selling 1001 really cool Web sites, this fun and informative book enables readers to take full advantage of the Web. More than a mere directory, it identifies and describes the best sites, guiding surfers to such innovations as VRML3-D and Java. Aside from downloads of Web browsers, Renehan points the way to free compilers and interpreters as well as free online access to major scientific journals
  19. Friedrich, M.; Schimkat, R.-D.; Küchlin, W.: Information retrieval in distributed environments based on context-aware, proactive documents (2002) 0.06
    0.06047188 = product of:
      0.24188752 = sum of:
        0.24188752 = weight(_text_:java in 4608) [ClassicSimilarity], result of:
          0.24188752 = score(doc=4608,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.5450528 = fieldWeight in 4608, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0546875 = fieldNorm(doc=4608)
      0.25 = coord(1/4)
    
    Abstract
    In this position paper we propose a document-centric middleware component called Living Documents to support context-aware information retrieval in distributed communities. A Living Document acts as a micro server for a document which contains computational services, a semi-structured knowledge repository to uniformly store and access context-related information, and finally the document's digital content. Our initial prototype of Living Documents is based an the concept of mobile agents and implemented in Java and XML.
  20. Hancock, B.; Giarlo, M.J.: Moving to XML : Latin texts XML conversion project at the Center for Electronic Texts in the Humanities (2001) 0.06
    0.06047188 = product of:
      0.24188752 = sum of:
        0.24188752 = weight(_text_:java in 5801) [ClassicSimilarity], result of:
          0.24188752 = score(doc=5801,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.5450528 = fieldWeight in 5801, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0546875 = fieldNorm(doc=5801)
      0.25 = coord(1/4)
    
    Abstract
    The delivery of documents on the Web has moved beyond the restrictions of the traditional Web markup language, HTML. HTML's static tags cannot deal with the variety of data formats now beginning to be exchanged between various entities, whether corporate or institutional. XML solves many of the problems by allowing arbitrary tags, which describe the content for a particular audience or group. At the Center for Electronic Texts in the Humanities the Latin texts of Lector Longinquus are being transformed to XML in readiness for the expected new standard. To allow existing browsers to render these texts, a Java program is used to transform the XML to HTML on the fly.

Languages

  • d 32
  • m 3
  • nl 1
  • More… Less…

Types

  • a 807
  • m 313
  • el 106
  • s 93
  • i 21
  • n 17
  • x 14
  • r 10
  • b 8
  • ? 1
  • v 1
  • More… Less…

Themes

Subjects

Classifications