Search (4850 results, page 3 of 243)

  • × year_i:[2000 TO 2010}
  1. Zhan, J.; Loh, H.T.: Using latent semantic indexing to improve the accuracy of document clustering (2007) 0.04
    0.0356988 = product of:
      0.1427952 = sum of:
        0.1427952 = weight(_text_:having in 1264) [ClassicSimilarity], result of:
          0.1427952 = score(doc=1264,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.39649835 = fieldWeight in 1264, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.046875 = fieldNorm(doc=1264)
      0.25 = coord(1/4)
    
    Abstract
    Document clustering is a significant research issue in information retrieval and text mining. Traditionally, most clustering methods were based on the vector space model which has a few limitations such as high dimensionality and weakness in handling synonymous and polysemous problems. Latent semantic indexing (LSI) is able to deal with such problems to some extent. Previous studies have shown that using LSI could reduce the time in clustering a large document set while having little effect on clustering accuracy. However, when conducting clustering upon a small document set, the accuracy is more concerned than efficiency. In this paper, we demonstrate that LSI can improve the clustering accuracy of a small document set and we also recommend the dimensions needed to achieve the best clustering performance.
  2. Chen, H.; Chung, W.; Qin, J.; Reid, E.; Sageman, M.; Weimann, G.: Uncovering the dark Web : a case study of Jihad on the Web (2008) 0.04
    0.0356988 = product of:
      0.1427952 = sum of:
        0.1427952 = weight(_text_:having in 2880) [ClassicSimilarity], result of:
          0.1427952 = score(doc=2880,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.39649835 = fieldWeight in 2880, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.046875 = fieldNorm(doc=2880)
      0.25 = coord(1/4)
    
    Abstract
    While the Web has become a worldwide platform for communication, terrorists share their ideology and communicate with members on the Dark Web - the reverse side of the Web used by terrorists. Currently, the problems of information overload and difficulty to obtain a comprehensive picture of terrorist activities hinder effective and efficient analysis of terrorist information on the Web. To improve understanding of terrorist activities, we have developed a novel methodology for collecting and analyzing Dark Web information. The methodology incorporates information collection, analysis, and visualization techniques, and exploits various Web information sources. We applied it to collecting and analyzing information of 39 Jihad Web sites and developed visualization of their site contents, relationships, and activity levels. An expert evaluation showed that the methodology is very useful and promising, having a high potential to assist in investigation and understanding of terrorist activities by producing results that could potentially help guide both policymaking and intelligence research.
  3. Robins, D.; Holmes, J.: Aesthetics and credibility in web site design (2008) 0.04
    0.0356988 = product of:
      0.1427952 = sum of:
        0.1427952 = weight(_text_:having in 3039) [ClassicSimilarity], result of:
          0.1427952 = score(doc=3039,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.39649835 = fieldWeight in 3039, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.046875 = fieldNorm(doc=3039)
      0.25 = coord(1/4)
    
    Abstract
    Web sites often provide the first impression of an organization. For many organizations, web sites are crucial to ensure sales or to procure services within. When a person opens a web site, the first impression is probably made in a few seconds, and the user will either stay or move on to the next site on the basis of many factors. One of the factors that may influence users to stay or go is the page aesthetics. Another reason may involve a user's judgment about the site's credibility. This study explores the possible link between page aesthetics and a user's judgment of the site's credibility. Our findings indicate that when the same content is presented using different levels of aesthetic treatment, the content with a higher aesthetic treatment was judged as having higher credibility. We call this the amelioration effect of visual design and aesthetics on content credibility. Our study suggests that this effect is operational within the first few seconds in which a user views a web page. Given the same content, a higher aesthetic treatment will increase perceived credibility.
  4. Bernstein, J.H.: Nonknowledge : the bibliographical organization of ignorance, stupidity, error, and unreason: Part Two (2009) 0.04
    0.0356988 = product of:
      0.1427952 = sum of:
        0.1427952 = weight(_text_:having in 3400) [ClassicSimilarity], result of:
          0.1427952 = score(doc=3400,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.39649835 = fieldWeight in 3400, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.046875 = fieldNorm(doc=3400)
      0.25 = coord(1/4)
    
    Abstract
    Starting with the Data-Information-Knowledge-Wisdom paradigm in information science, it is possible to derive a model of the opposite of knowledge having hierarchical qualities. A range of counterpoints to concepts in the knowledge hierarchy can be identified and ascribed the overall term "nonknowledge." This model creates a conceptual framework for understanding the connections between topics such as error, ignorance, stupidity, folly, popular misconceptions, and unreason, by locating them as levels or phases of nonknowledge. The concept of nonknowledge links heretofore disconnected discourses on these individual topics by philosophers, psychologists, historians, sociologists, satirists, and others. Subject headings provide access to the categories of nonknowledge, but confusion remains due to the general failure of cataloging and classification to differentiate between works about nonknowledge and examples of nonknowledge.
  5. Levergood, B.; Farrenkopf, S.; Frasnelli, E.: ¬The specification of the language of the field and interoperability : cross-language access to catalogues and online libraries (CACAO) (2008) 0.04
    0.0356988 = product of:
      0.1427952 = sum of:
        0.1427952 = weight(_text_:having in 3646) [ClassicSimilarity], result of:
          0.1427952 = score(doc=3646,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.39649835 = fieldWeight in 3646, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.046875 = fieldNorm(doc=3646)
      0.25 = coord(1/4)
    
    Abstract
    The CACAO Project (Cross-language Access to Catalogues and Online Libraries) has been designed to implement natural language processing and cross-language information retrieval techniques to provide cross-language access to information in libraries, a critical issue in the linguistically diverse European Union. This project report addresses two metadata-related challenges for the library community in this context: "false friends" (identical words having different meanings in different languages) and term ambiguity. The possible solutions involve enriching the metadata with attributes specifying language or the source authority file, or associating potential search terms to classes in a classification system. The European Library will evaluate an early implementation of this work in late 2008.
  6. Chen, Y.-H.; Germain, C.A.; Yang, H.: ¬An exploration into the practices of library Web usability in ARL academic libraries (2009) 0.04
    0.0356988 = product of:
      0.1427952 = sum of:
        0.1427952 = weight(_text_:having in 3798) [ClassicSimilarity], result of:
          0.1427952 = score(doc=3798,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.39649835 = fieldWeight in 3798, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.046875 = fieldNorm(doc=3798)
      0.25 = coord(1/4)
    
    Abstract
    A survey of the 113 academic libraries of the Association of Research Libraries (ARL) was administered to investigate whether Web usability Policies/Standards/Guidelines (PSGs) are in place, the levels of difficulty surrounding implementation, the impact of PSGs on actual usability practice, e.g., testing, resources, etc., and the relationship between ARL ranking and usability practice or PSGs. The response rate was over 74%. Results show that 25 (30%) libraries have PSGs dedicated to Web usability. Seventy-one (85%) libraries have conducted usability testing on their main Web sites, online public access catalogs (OPAC), or lower-level pages. Nevertheless, only seven libraries performed iterative testing of these platforms at pre-, during, and post-design stages. Statistical analysis indicates that having PSGs does not affect the amount of usability testing performed or the resources available for library Web usability initiatives. In addition, ARL ranking has no or little impact on PSGs, testing, or resources.
  7. Greenberg, J.: Automatic query expansion via lexical-semantic relationships (2001) 0.03
    0.029749004 = product of:
      0.11899602 = sum of:
        0.11899602 = weight(_text_:having in 6703) [ClassicSimilarity], result of:
          0.11899602 = score(doc=6703,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.3304153 = fieldWeight in 6703, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.0390625 = fieldNorm(doc=6703)
      0.25 = coord(1/4)
    
    Abstract
    Structured thesauri encode equivalent, hierarchical, and associative relationships and have been developed as indexing/retrieval tools. Despite the fact that these tools provide a rich semantic network of vocabulary terms, they are seldom employed for automatic query expansion (QE) activities. This article reports on an experiment that examined whether thesaurus terms, related to query in a specified semantic way (as synonyms and partial-synonyms (SYNs), narrower terms (NTs), related terms (RTs), and broader terms (BTs)), could be identified as having a more positive impact on retrieval effectiveness when added to a query through automatic QE. The research found that automatic QE via SYNs and NTs increased relative recall with a decline in precision that was not statistically significant, and that automatic QE via RTs and BTs increased relative recall with a decline in precision that was statistically significant. Recallbased and a precision-based ranking orders for automatic QE via semantically encoded thesauri terminology were identified. Mapping results found between enduser query terms and the ProQuest Controlled Vocabulary (1997) (the thesaurus used in this study) are reported, and future research foci related to the investigation are discussed
  8. Andersen, J.: Written knowledge : a literary perspective on indexing theory (2000) 0.03
    0.029749004 = product of:
      0.11899602 = sum of:
        0.11899602 = weight(_text_:having in 87) [ClassicSimilarity], result of:
          0.11899602 = score(doc=87,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.3304153 = fieldWeight in 87, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.0390625 = fieldNorm(doc=87)
      0.25 = coord(1/4)
    
    Abstract
    The general concern of Bazerman's book Shaping Written Knowledge. The Genre and Activity of the Experimental Article in Science is written knowledge as it is produced by the academy. Bazerman discusses in particular the rhetoric, communicative, and epistemological issues of written knowledge. The article discusses these themes in a library and information science (LIS) perspective in terms of their implications for LIS research. For several reasons, it is argued that this way of scrutinizing into written knowledge ought to be of special interest to LIS research. As an example of a particular field of research in LIS, the article discusses the relationship between indexing theory and written knowledge. Bazerman analyzes written knowledge from a literary point of view. Among other things, it is argued that indexing theory can be seen as part of literary theory in that some of the questions raised by the latter are also raised in indexing theory. Furthermore, it is put forward that the indexer can be considered an author. The indexer produces a text, the document representation, which is the text the user actually meets in the first place. That way, the producer of a document representation is to some extent responsible for the quality of the documents indexed. Having discussed this relationship between written knowledge and LIS research in general and indexing theory in particular, it is concluded that LIS research ought to head toward more humanistic oriented research traditions, if the line of research presented by Bazerman should be considered useful for LIS
  9. Fraser, L.; Locatis, C.: Effects of link annotations on search performance in layered and unlayered hierarchically organized information spaces (2001) 0.03
    0.029749004 = product of:
      0.11899602 = sum of:
        0.11899602 = weight(_text_:having in 937) [ClassicSimilarity], result of:
          0.11899602 = score(doc=937,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.3304153 = fieldWeight in 937, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.0390625 = fieldNorm(doc=937)
      0.25 = coord(1/4)
    
    Abstract
    The effects of link annotations on user search performance in hypertext environments having deep (layered) and shallow link structures were investigated in this study. Four environments were tested-layered-annotated, layered-unannotated, shallow-annotated, and shallow-unannotated. A single document was divided into 48 sections, and layered and unlayered versions were created. Additional versions were created by adding annotations to the links in the layered and unlayered versions. Subjects were given three queries of varying difficulty and then asked to find the answers to the queries that were contained within the hypertext environment to which they were randomly assigned. Correspondence between the wording links and queries was used to define difficulty level. The results of the study confirmed previous research that shallow link structures are better than deep (layered) link structures. Annotations had virtually no effect on the search performance of the subjects. The subjects performed similarly in the annotated and unannotated environments, regardless of whether the link structures were shallow or deep. An analysis of question difficulty suggests that the wording in links has primacy over the wording in annotations in influencing user search behavior.
  10. Dong-Geun, O.; Ji-Suk, Y.: Suggesting an option for DDC class religion (200) for nations in which religious diversity predominates (2001) 0.03
    0.029749004 = product of:
      0.11899602 = sum of:
        0.11899602 = weight(_text_:having in 956) [ClassicSimilarity], result of:
          0.11899602 = score(doc=956,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.3304153 = fieldWeight in 956, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.0390625 = fieldNorm(doc=956)
      0.25 = coord(1/4)
    
    Abstract
    This study makes recommendations to adapt and expand the class Religion (200) of the Dewey Decimal Classification (DDC) for libraries in nations, particularly Korea, where religious diversity predominates. It contracts the classes on Christianity, 220-280, into one class, 220, and allows for local Korean emphasis so that religions of Asian, and especially East Asian, origin are given preferred treatment. It relocates Buddhism and other religions of Asian origin into separate divisions, 230 and 240, respectively. The recommendations maintain the basic system and order of DDC. Whenever possible, they adhere to the Korean Decimal Classification in the arrangement of religions of Asian origin. Because the options for religions other than Christianity in DDC are to give preferred treatment to a specific religion, this article tries to generalize the idea to suggest an option in DDC as Option F for "> 292-299 Religions other than Christianity", for other nations in the world having similar situations. It says: "Contract the Bible and Christianity in 220-280 into 220, and then class other religions to give preferred treatment in 230-280, respectively, following the order in DDC. In that case, the number "2" should be inserted in the second digit of the Bible and Christianity, e.g., the Bible 222, Jesus Christ & his family 223.2, and Puritanism 228.59."
  11. Johnson, E.H.: Objects for distributed heterogeneous information retrieval (2000) 0.03
    0.029749004 = product of:
      0.11899602 = sum of:
        0.11899602 = weight(_text_:having in 959) [ClassicSimilarity], result of:
          0.11899602 = score(doc=959,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.3304153 = fieldWeight in 959, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.0390625 = fieldNorm(doc=959)
      0.25 = coord(1/4)
    
    Abstract
    The success of the World Wide Web Shows that we can access, search, and retrieve information from globally distributed databases. lf a database, such as a library catalog, has some sort of Web-based front end, we can type its URL into a Web browser and use its HTML-based forms to search for items in that database. Depending an how well the query conforms to the database content, how the search engine interprets the query, and how the server formats the results into HTML, we might actually find something usable. While the first two issues depend an ourselves and the server, an the Web the latter falls to the mercy of HTML, which we all know as a great destroyer of information because it codes for display but not for content description. When looking at an HTML-formatted display, we must depend an our own interpretation to recognize such entities as author names, titles, and subject identifiers. The Web browser can do nothing but display the information. lf we want some other view of the result, such as sorting the records by date (provided it offers such an option to begin with), the server must do it. This makes poor use of the computing power we have at the desktop (or even laptop), which, unless it involves retrieving more records, could easily do the result Set manipulation that we currently send back to the server. Despite having personal computers wich immense computational power, as far as information retrieval goes, we still essentially use them as dumb terminals.
  12. Special topic issue: XML (2002) 0.03
    0.029749004 = product of:
      0.11899602 = sum of:
        0.11899602 = weight(_text_:having in 1458) [ClassicSimilarity], result of:
          0.11899602 = score(doc=1458,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.3304153 = fieldWeight in 1458, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1458)
      0.25 = coord(1/4)
    
    Abstract
    The Web has driven many developments related to information retrieval in the past 5 years. One of the recent waves is due to XML (eXtensible Markup Language), a meta-language derived from SGML and designed for data representation and exchange on the Internet. XML provides support for marking (tagging) text for many purposes such as structure, semantics, layout, etc. It is believed that it will become a universal format not only for business-to-business applications but also for knowledge and information management. As a result, it has become crucial to address the question of how large collections of XML documents can be sorted and retrieved efficiently and effectively. Furthermore, it is necessary to develop a suitable query language for XML documents. Last year the World Wide Web Consortium released a query language proposal for XML data known as XQuery. XQuery is the combination of many proposals and previous languages, some of which are covered in this issue. XQuery is the natural meeting point of the database and information retrieval communities, because it requires that SQL-like queries as well as text retrieval queries must be answered correctly and efficiently. Hence, having a special issue on XML and information retrieval is timely and almost compulsory.
  13. Fugmann, R.: Obstacles to progress in mechanized subject access and the necessity of a paradigm change (2000) 0.03
    0.029749004 = product of:
      0.11899602 = sum of:
        0.11899602 = weight(_text_:having in 2182) [ClassicSimilarity], result of:
          0.11899602 = score(doc=2182,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.3304153 = fieldWeight in 2182, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.0390625 = fieldNorm(doc=2182)
      0.25 = coord(1/4)
    
    Abstract
    Contemporary information systems, both the private and the commercially available ones, have often been blamed for their low effectiveness in terms of precision and recall, especially when they have reached considerable size with respect to file volume and use frequency (see, for example, Belkin, 1980; Blair, 1996, p.19; Desai, 1997; Drabenstott, 1996; Knorz, 1998). Saracevic (1989), after having reviewed the contemporary design of online subject access, calls "for radically different design principles and implementation" (p. 107). Van Rijsbergen (1990) writes: "The keywords approach with statistical techniques has reached its theoretical limit and further attempts for improvement are considered a waste of time" (p. 111). Lancaster (1992) deplores that very little really significant literature an subject indexing has been published in the last thirty or so years. In her preface to the Proceedings of the Sixth International Study Conference an Classification Research in 1997, Mcllwaine (1997) writes, "many were surprised to find that the problems with which they wrestle today are not greatly different from those that have been occupying the minds of specialists in the field for over a generation, and probably a great deal longer" (p. v).
  14. Hars, A.: From publishing to knowledge networks : reinventing online knowledge infrastructures (2003) 0.03
    0.029749004 = product of:
      0.11899602 = sum of:
        0.11899602 = weight(_text_:having in 2634) [ClassicSimilarity], result of:
          0.11899602 = score(doc=2634,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.3304153 = fieldWeight in 2634, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.0390625 = fieldNorm(doc=2634)
      0.25 = coord(1/4)
    
    Abstract
    Today's publishing infrastructure is rapidly changing. As electronic journals, digital libraries, collaboratories, logic servers, and other knowledge infrastructures emerge an the internet, the key aspects of this transformation need to be identified. Knowledge is becoming increasingly dynamic and integrated. Instead of writing self-contained articles, authors are turning to the new practice of embedding their findings into dynamic networks of knowledge. Here, the author details the implications that this transformation is having an the creation, dissemination and organization of academic knowledge. The author Shows that many established publishing principles need to be given up in order to facilitate this transformation. The text provides valuable insights for knowledge managers, designers of internet-based knowledge infrastructures, and professionals in the publishing industry. Researchers will find the scenarios and implications for research processes stimulating and thought-provoking.
  15. Polanco, X.: Clusters, graphs, and networks for analyzing Internet-Web-supported communication within a virtual community (2003) 0.03
    0.029749004 = product of:
      0.11899602 = sum of:
        0.11899602 = weight(_text_:having in 3737) [ClassicSimilarity], result of:
          0.11899602 = score(doc=3737,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.3304153 = fieldWeight in 3737, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.0390625 = fieldNorm(doc=3737)
      0.25 = coord(1/4)
    
    Abstract
    The proposal is to use clusters, graphs and networks as models in order to analyse the Web structure. Clusters, graphs and networks provide knowledge representation and organization. Clusters were generated by co-site analysis. The sample is a set of academic Web sites from the countries belonging to the European Union. These clusters are here revisited from the point of view of graph theory and social network analysis. This is a quantitative and structural analysis. In fact, the Internet is a computer network that connects people and organizations. Thus we may consider it to be a social network. The set of Web academic sites represents an empirical social network, and is viewed as a virtual community. The network structural properties are here analysed applying together cluster analysis, graph theory and social network analysis. This is a work having taken place in the EICSTES project. EICSTES means European Indicators, Cyberspace, and the Science-Technology-Economy System. It is a research project supported by the Fifth Framework Program of R&D of the European Commission (IST-1999-20350)
  16. Yager, R.R.: Knowledge trees and protoforms in question-answering systems (2006) 0.03
    0.029749004 = product of:
      0.11899602 = sum of:
        0.11899602 = weight(_text_:having in 281) [ClassicSimilarity], result of:
          0.11899602 = score(doc=281,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.3304153 = fieldWeight in 281, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.0390625 = fieldNorm(doc=281)
      0.25 = coord(1/4)
    
    Abstract
    We point out that question-answering systems differ from other information-seeking applications, such as search engines, by having a deduction capability, an ability to answer questions by a synthesis of information residing in different parts of its knowledge base. This capability requires appropriate representation of various types of human knowledge, rules for locally manipulating this knowledge, and a framework for providing a global plan for appropriately mobilizing the information in the knowledge to address the question posed. In this article we suggest tools to provide these capabilities. We describe how the fuzzy set-based theory of approximate reasoning can aid in the process of representing knowledge. We discuss how protoforms can be used to aid in deduction and local manipulation of knowledge. The idea of a knowledge tree is introduced to provide a global framework for mobilizing the knowledge base in response to a query. We look at some types of commonsense and default knowledge. This requires us to address the complexity of the nonmonotonicity that these types of knowledge often display. We also briefly discuss the role that Dempster-Shafer structures can play in representing knowledge.
  17. Subrahmanyam, B.: Library of Congress Classification numbers : issues of consistency and their implications for union catalogs (2006) 0.03
    0.029749004 = product of:
      0.11899602 = sum of:
        0.11899602 = weight(_text_:having in 784) [ClassicSimilarity], result of:
          0.11899602 = score(doc=784,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.3304153 = fieldWeight in 784, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.0390625 = fieldNorm(doc=784)
      0.25 = coord(1/4)
    
    Abstract
    This study examined Library of Congress Classification (LCC)-based class numbers assigned to a representative sample of 200 titles in 52 American library systems to determine the level of consistency within and across those systems. The results showed that under the condition that a library system has a title, the probability of that title having the same LCC-based class number across library systems is greater than 85 percent. An examination of 121 titles displaying variations in class numbers among library systems showed certain titles (for example, multi-foci titles, titles in series, bibliographies, and fiction) lend themselves to alternate class numbers. Others were assigned variant numbers either due to latitude in the schedules or for reasons that cannot be pinpointed. With increasing dependence on copy cataloging, the size of such variations may continue to decrease. As the preferred class number with its alternates represents a title more fully than just the preferred class number, this paper argues for continued use of alternates by library systems and for finding a method to link alternate class numbers to preferred class numbers for enriched subject access through local and union catalogs.
  18. Aula, A.; Nordhausen, K.: Modeling successful performance in Web searching (2006) 0.03
    0.029749004 = product of:
      0.11899602 = sum of:
        0.11899602 = weight(_text_:having in 117) [ClassicSimilarity], result of:
          0.11899602 = score(doc=117,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.3304153 = fieldWeight in 117, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.0390625 = fieldNorm(doc=117)
      0.25 = coord(1/4)
    
    Abstract
    Several previous studies have measured differences in the information search success of novices and experts. However, the definitions of novices and experts have varied greatly between the studies, and so have the measures used for search success. Instead of dividing the searchers into different groups based on their expertise, we chose to model search success with task completion speed, TCS. Towards this goal, 22 participants performed three fact-finding tasks and two broader tasks in an observational user study. In our model, there were two variables related to the Web experience of the participants. Other variables included, for example, the speed of query iteration, the length of the queries, the proportion of precise queries, and the speed of evaluating result documents. Our results showed that the variables related to Web experience had expected effects on TCS. The increase in the years of Web use was related to improvement in TCS in the broader tasks, whereas the less frequent Web use was related to a decrease in TCS in the fact-finding tasks. Other variables having significant effects on TCS in either of the task types were the speed of composing queries, the average number of query terms per query, the proportion of precise queries, and the participants' own evaluation of their search skills. In addition to the statistical models, we present several qualitative findings of the participants' search strategies. These results give valuable insight into the successful strategies in Web search beyond the previous knowledge of the expert-novice differences.
  19. Lundy, M.W.: Use and perception of the DCRB Core standard (2003) 0.03
    0.029749004 = product of:
      0.11899602 = sum of:
        0.11899602 = weight(_text_:having in 278) [ClassicSimilarity], result of:
          0.11899602 = score(doc=278,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.3304153 = fieldWeight in 278, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.0390625 = fieldNorm(doc=278)
      0.25 = coord(1/4)
    
    Abstract
    In January 1999, the Program for Cooperative Cataloging approved the core bibliographic standard for rare books, called the DCRB Core standard. Like the other core standards, the DCRB Core provides the framework within which catalogers can create bibliographic records that are less than full, but are as reliable as full-level records in description and authorized headings. In the three years since its approval, there is little evidence that the standard has been widely used. This study reports the results of a survey sent to forty-three participants who indicated in a preliminary query that they do use the DCRB Core or that they have made the decision not to use it. In the thirty-seven surveys that were returned, only about 16% of the respondents said they have used the standard to create bibliographic records for their rare books. The libraries that do not use the core standard find it inferior or lacking in a number of ways. Several of those libraries, however, are planning to use the standard in the future or are seriously planning to investigate using it. Such intent may indicate that the time is approaching when more libraries will find reasons to implement the standard. One impetus may come from the findings of a recent survey of the special collections departments of member libraries of the Association of Research Libraries that emphasize the size of the backlogs in those departments. If faster accessibility to specific portions of the backlogs would benefit users more than having fulllevel cataloging, application of the DCRB Core standard could facilitate reducing those backlogs.
  20. Abd Manaf, Z.; Abdul Rahman, R.: Examining the quality of National Library of Malaysia (NLM) cataloguing in publication (CIP) records (2006) 0.03
    0.029749004 = product of:
      0.11899602 = sum of:
        0.11899602 = weight(_text_:having in 747) [ClassicSimilarity], result of:
          0.11899602 = score(doc=747,freq=2.0), product of:
            0.36014074 = queryWeight, product of:
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.060212567 = queryNorm
            0.3304153 = fieldWeight in 747, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              5.981156 = idf(docFreq=304, maxDocs=44421)
              0.0390625 = fieldNorm(doc=747)
      0.25 = coord(1/4)
    
    Abstract
    Purpose - This study attempts to compare the National Library of Malaysia (NLM)'s cataloguing in publication (CIP) records with their permanent records in the NLM's online public access catalogue (OPAC) database. Design/methodology/approach - The study compares the description of records in the NLM CIP versus NLM OPAC in terms of similarities and differences in the bibliographic elements for each record. The study selected the NLM's CIP records from January to April 2000, allowing more than 5 years for publishers to publish the registered CIP titles and also the NLM to make the records available in their OPAC database. Findings - The findings indicate that libraries that use NLM's CIP records in their copy cataloguing activity should not rely 100 per cent on the records. Verifications and modifications need to be carried out to ensure accuracy and quality of their catalogue records. Research limitations/implications - That Malaysian experience confirms research carried out previously elsewhere. Practical implications - The findings of the study are significant in terms of giving data about whether NLM's CIP records are trustworthy sources for the practical task of copy cataloguing or not. Originality/value - Although comparison studies of the accuracy and consistency of CIP records versus OPAC records are commonplace, having been conducted extensively since the 1980s in a variety of contexts, such a comparison study of CIP records versus OPAC records has never been conducted in Malaysia. With the increased usage of web-based OPACs as one source option for copy cataloguing, it is imperative to ensure the records copied are consistent and accurate to avoid post-alteration of the bibliographic description.

Languages

  • d 4222
  • e 586
  • m 11
  • es 2
  • f 2
  • s 2
  • el 1
  • ru 1
  • More… Less…

Types

  • a 3616
  • m 820
  • el 228
  • x 211
  • s 190
  • i 47
  • r 29
  • n 8
  • b 7
  • l 5
  • More… Less…

Themes

Subjects

Classifications