Search (13355 results, page 10 of 668)

  1. Aldous, K.J.: ¬A system for the automatic retrieval of information from a specialist database (1996) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 4146) [ClassicSimilarity], result of:
          0.18536724 = score(doc=4146,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 4146, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=4146)
      0.25 = coord(1/4)
    
    Abstract
    Accessing useful information from a complex database requires knowledge of the structure of the database and an understanding of the methods of information retrieval. A means of overcoming this knowledge barrier to the use of narrow domain databases is proposed in which the user is required to enter only a series of terms which identify the required material. Describes a method which classifies terms according to their meaning in the context of the database and which uses this classification to access and execute models of code stored in the database to effect retrieval. Presents an implementation of the method using a database of technical information on the nature and use of fungicides. Initial results of trials with potential users indicate that the system can produce relevant resposes to queries expressed in this style. Since the code modules are part of the database, extensions may be easily implemented to handle most queries which users are likely to pose
  2. Gottlob, G.; Shrefl, M.; Rock, B.: Extending object-oriented systems with roles (1996) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 6725) [ClassicSimilarity], result of:
          0.18536724 = score(doc=6725,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 6725, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=6725)
      0.25 = coord(1/4)
    
    Abstract
    Shows how class-based object-oriented systems can be extended to handle evolving objects. Class hierarchies are complemented by role hierarchies, whose nodes represent role types that an object classified in the root may take on. At any point in time, an entity is represented by an instance of the root and an instance of every role type whose role it curently plays. In a natural way, the approach extends traditional object oriented concepts, such as classification, object identity, specialization, inheritance, and polymorphism in a natural way. Demonstartes the practicability of the approach by an implementation in Smalltalk. Roles can be provided in Smalltalk by adding a few classes. There is no need to modify the semantics of Smalltalk itself. Role hierarchies are mapped transparently onto ordinary classes. The presented implementation can easily be ported to object-oriented database programming languages based on Smalltalk such as Gemstone's OPAL
  3. Mehrotra, R.: Content-based image modeling and retrieval (1997) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 761) [ClassicSimilarity], result of:
          0.18536724 = score(doc=761,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 761, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=761)
      0.25 = coord(1/4)
    
    Abstract
    Standard databases handle an abstracted world pertinent to specific problems in terms of alphanumeric data and respond to semantics defined by users and known to the databases. This conventional approach is not well suited to image databases, where the desired information and semantics need to be automatically or semi automatically extracted and and appropriately modelled to facilitate content based information retrieval and data manioulation. Discusses the key issues in content based image data modelling and information retrieval with particular reference to the MUlti-SEmantic Unstructured data Model (MUSEUM) system which is capable of dealing with image representatios of 3 dimensional shapes
  4. Darnard, D.T.; Ide, N.M.: ¬The Text Encoding Initiative : flexible and extensible document encoding (1997) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 1388) [ClassicSimilarity], result of:
          0.18536724 = score(doc=1388,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 1388, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=1388)
      0.25 = coord(1/4)
    
    Abstract
    The Text Encoding Initiative is an international collaboration aimed at producing an common encoding scheme for complex texts. The diversity of the texts used by members of the communities served by the project led to a large specification, but the specification is structured to facilitate understanding and use. The requirement for generality is sometimes in tension with the requirement to handle specialized text types. The texts that are encoded often can be viewed or interpreted in severla different ways. While many electronic documents can be encoded in very simple ways, some document and some users will tax the limits of any fixed scheme, so a flexible extensible encoding is required to support research and to facilitate the reuse of texts
  5. Hmeidi, I.; Kanaan, G.; Evens, M.: Design and implementation of automatic indexing for information retrieval with Arabic documents (1997) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 2660) [ClassicSimilarity], result of:
          0.18536724 = score(doc=2660,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 2660, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=2660)
      0.25 = coord(1/4)
    
    Abstract
    A corpus of 242 abstracts of Arabic documents on computer science and information systems using the Proceedings of the Saudi Arabian National Conferences as a source was put together. Reports on the design and building of an automatic information retrieval system from scratch to handle Arabic data. Both automatic and manual indexing techniques were implemented. Experiments using measures of recall and precision has demonstrated that automatic indexing is at least as effective as manual indexing and more effective in some cases. Automatic indexing is both cheaper and faster. Results suggests that a wider coverage of the literature can be achieved with less money and produce as good results as with manual indexing. Compares the retrieval results using words as index terms versus stems and roots, and confirms the results obtained by Al-Kharashi and Abu-Salem with smaller corpora that root indexing is more effective than word indexing
  6. Verhoeven, A.A. H.; Boerma, E.J.; Jong, M.-D.: Use of information sources by family physicians : a literature survey (1995) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 3796) [ClassicSimilarity], result of:
          0.18536724 = score(doc=3796,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 3796, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=3796)
      0.25 = coord(1/4)
    
    Abstract
    State of the art review of the use of information sources by family doctors is important for both practical and theoretical reasons. Analysis of the ways in which family doctors handle information may point to opportunities for improvement. Such efforts lead to improvements in the methodology of literature research in general. Reports on a review of the literature on information use by family doctors. 11 relevant research publications were found. The data showed that family doctors used colleagues most often as information sources followed by periodicals and books. This outcome corresponded with results found for other professions. Several factors influenced the use of information sources by family doctors, including the physical, functional and intellectual accessibility of the sources; the doctor's age; participation by the doctor in research or education; the social context of the doctor; practice characteristics; and the stage of the informatio gathering process. The publications studied suggested ways to improve information gathering in the areas of computerization, education, library organization, and periodical articles
  7. Wong, M.L.; Leung, K.S.; Cheng, J.C.Y.: Discovering knowledge from noisy databases using genetic programming (2000) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 5863) [ClassicSimilarity], result of:
          0.18536724 = score(doc=5863,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 5863, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=5863)
      0.25 = coord(1/4)
    
    Abstract
    In data mining, we emphasize the need for learning from huge, incomplete, and imperfect data sets. To handle noise in the problem domain, existing learning systems avoid overfitting the imperfect training examples by excluding insignificant patterns. The problem is that these systems use a limiting attribute-value language for representing the training examples and the induced knowledge. Moreover, some important patterns are ignored because they are statistically insignificant. In this article, we present a framework that combines genetic programming and inductive logic programming to induce knowledge represented in various knowledge representation formalisms from noisy databases (LOGENPRO). Moreover, the system is applied to one real-life medical database. The knowledge discovered provides insights to and allows better understanding of the medical domains
  8. Information technology for knowledge management (1998) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 1690) [ClassicSimilarity], result of:
          0.18536724 = score(doc=1690,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 1690, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=1690)
      0.25 = coord(1/4)
    
    Abstract
    The value of an organization is given not just by its tangible assets but also by the knowledge embodied in its employees and in its internal and external structures. Explicit knowledge can be shared as information and is easy to handle. Tacit knowledge has been neglected by effectiveness-orientied management techniques but is crucial for both the operational efficiency and the core competence of an organization. This book provided a survey of the use of information technology for knowledge management, and its chapters present specific research on how technologies such as computer-supported cooperative work (CSCW), workflow, and groupware can support the creation and sharing of knowledge in organizations
  9. Ferret, O.; Grau, B.; Hurault-Plantet, M.; Illouz, G.; Jacquemin, C.; Monceaux, L.; Robba, I.; Vilnat, A.: How NLP can improve question answering (2002) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 2850) [ClassicSimilarity], result of:
          0.18536724 = score(doc=2850,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 2850, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=2850)
      0.25 = coord(1/4)
    
    Abstract
    Answering open-domain factual questions requires Natural Language processing for refining document selection and answer identification. With our system QALC, we have participated in the Question Answering track of the TREC8, TREC9 and TREC10 evaluations. QALC performs an analysis of documents relying an multiword term searches and their linguistic variation both to minimize the number of documents selected and to provide additional clues when comparing question and sentence representations. This comparison process also makes use of the results of a syntactic parsing of the questions and Named Entity recognition functionalities. Answer extraction relies an the application of syntactic patterns chosen according to the kind of information that is sought, and categorized depending an the syntactic form of the question. These patterns allow QALC to handle nicely linguistic variations at the answer level.
  10. Whitmire, E.: ¬The relationship between undergraduates' epistemological beliefs, reflective judgment, and their information-seeking behavior (2004) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 3561) [ClassicSimilarity], result of:
          0.18536724 = score(doc=3561,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 3561, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=3561)
      0.25 = coord(1/4)
    
    Abstract
    During the fall 2001 semester 15 first-year undergraduates were interviewed about their information-seeking behavior. Undergraduates completed a short-answer questionnaire, the Measure of Epistemological Reflection, measuring their epistemological beliefs and searched the Web and an online public access catalog using tasks from the Reflective Judgment Interview that assessed their reflective judgment level. Undergraduates talked aloud while searching digital environments about the decisions they were making about the information they encountered while transaction analyses software (Lotus ScreenCam) recorded both their search moves and their decision-making through verbal protocol analysis. Analyses included examining the relationship between undergraduates' epistemological beliefs and reflective judgment and how they searched for information in these digital environments. Results indicated that there was a relationship between epistemological beliefs and reflective judgment and information-seeking behavior. Undergraduates' at higher stages of epistemological development exhibited the ability to handle conflicting information sources and to recognize authoritative information sources.
  11. Spink, A.; Park, M.: Information and non-information multitasking interplay (2005) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 5330) [ClassicSimilarity], result of:
          0.18536724 = score(doc=5330,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 5330, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=5330)
      0.25 = coord(1/4)
    
    Abstract
    Purpose - During multitasking, humans handle multiple tasks through task switching or engage in multitasking information behaviors. For example, a user switches between seeking new kitchen information and medical information. Recent studies provide insights these complex multitasking human information behaviors (HIB). However, limited studies have examined the interplay between information and non-information tasks. Design/methodology/approach - The goal of the paper was to examine the interplay of information and non-information task behaviors. Findings - This paper explores and speculates on a new direction in HIB research. The nature of HIB as a multitasking activity including the interplay of information and non-information behavior tasks, and the relation between multitasking information behavior to cognitive style and individual differences, is discussed. A model of multitasking between information and non-information behavior tasks is proposed. Practical implications/limitations - Multitasking information behavior models should include the interplay of information and non-information tasks, and individual differences and cognitive styles. Originality/value - The paper is the first information science theoretical examination of the interplay between information and non-information tasks.
  12. Collins, W.P.: Classification of materials on the Bahá'í religion : expansion of Library of Congress BP300-395 (1988) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 513) [ClassicSimilarity], result of:
          0.18536724 = score(doc=513,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 513, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=513)
      0.25 = coord(1/4)
    
    Abstract
    When the Bahá'í World Centre Library was organized on professional principles in 1977, there was no system adequate to handle the classification of a large specialized collection on the Bahá'í Faith. The librarian, in consultation with knowledgeable Bahá'í is and Bahá'í librarians, began a staged development of a classification scheme that could be adapted to the uses of institutions with significant collections of Bahá'í material. Following the development of two preliminary expansions of Dewey and Library of Congress systems, and study of the types and subject matter of materials being produced, a reasonably complete classification for the current subject areas covered in Bahá'í publications was developed. This classification should be of benefit to academic libraries planning large acquisitions of Bahá'í materials, and particularly to the growing numbers of developing Bahá'í libraries.
  13. Hoban, M.S.: Sound recording cataloging : a practical approach (1990) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 625) [ClassicSimilarity], result of:
          0.18536724 = score(doc=625,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 625, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=625)
      0.25 = coord(1/4)
    
    Abstract
    The cataloging of music sound recordings is both challenging and interesting. As the technologies used to produce sound recordings change, these changes must be reflected in both cataloging theory and practice. Three formats: analog disc, cassette tape, and compact disk, all of which are readily available on the market, present special challenges to catalogers who must consider what might be the most effective way of handling these materials following the AACR2 cataloging rules and interpretations from Library of Congress. This paper examines the actual cataloging of those formats as done by several institutions and raises questions such as how to handle these materials in ways which will eliminate redundancy and increase efficiency in the practice of cataloging. Finally, an alternative approach, drawing on AACR2 practice in other areas, is suggested.
  14. Stapleton, M.; Adams, M.: Faceted categorisation for the corporate desktop : visualisation and interaction using metadata to enhance user experience (2007) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 1718) [ClassicSimilarity], result of:
          0.18536724 = score(doc=1718,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 1718, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=1718)
      0.25 = coord(1/4)
    
    Abstract
    Mark Stapleton and Matt Adamson began their presentation by describing how Dow Jones' Factiva range of information services processed an average of 170,000 documents every day, drawn from over 10,000 sources in 22 languages. These documents are categorized within five facets: Company, Subject, Industry, Region and Language. The digital feeds received from information providers undergo a series of processing stages, initially to prepare them for automatic categorization and then to format them ready for distribution. The categorization stage is able to handle 98% of documents automatically, the remaining 2% requiring some form of human intervention. Depending on the source, categorization can involve any combination of 'Autocoding', 'Dictionary-based Categorizing', 'Rules-based Coding' or 'Manual Coding'
  15. Spink, A.; Park, M.; Koshman, S.: Factors affecting assigned information problem ordering during Web search : an exploratory study (2006) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 1991) [ClassicSimilarity], result of:
          0.18536724 = score(doc=1991,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 1991, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=1991)
      0.25 = coord(1/4)
    
    Abstract
    Multitasking is the human ability to handle the demands of multiple tasks. Multitasking behavior involves the ordering of multiple tasks and switching between tasks. People often multitask when using information retrieval (IR) technologies as they seek information on more than one information problem over single or multiple search episodes. However, limited studies have examined how people order their information problems, especially during their Web search engine interaction. The aim of our exploratory study was to investigate assigned information problem ordering by forty (40) study participants engaged in Web search. Findings suggest that assigned information problem ordering was influenced by the following factors, including personal interest, problem knowledge, perceived level of information available on the Web, ease of finding information, level of importance and seeking information on information problems in order from general to specific. Personal interest and problem knowledge were the major factors during assigned information problem ordering. Implications of the findings and further research are discussed. The relationship between information problem ordering and gratification theory is an important area for further exploration.
  16. Käki, M.; Aula, A.: Controlling the complexity in comparing search user interfaces via user studies (2008) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 3024) [ClassicSimilarity], result of:
          0.18536724 = score(doc=3024,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 3024, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=3024)
      0.25 = coord(1/4)
    
    Abstract
    Over time, researchers have acknowledged the importance of understanding the users' strategies in the design of search systems. However, when involving users in the comparison of search systems, methodological challenges still exist as researchers are pondering on how to handle the variability that human participants bring to the comparisons. This paper present methods for controlling the complexity of user-centered evaluations of search user interfaces through within-subjects designs, balanced task sets, time limitations, pre-formulated queries, cached result pages, and through limiting the users' access to result documents. Additionally, we will present our experiences in using three measures - search speed, qualified search speed, and immediate accuracy - to facilitate the comparison of different search systems over studies.
  17. Zhang, Y.: Complex adaptive filtering user profile using graphical models (2008) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 3445) [ClassicSimilarity], result of:
          0.18536724 = score(doc=3445,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 3445, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=3445)
      0.25 = coord(1/4)
    
    Abstract
    This article explores how to develop complex data driven user models that go beyond the bag of words model and topical relevance. We propose to learn from rich user specific information and to satisfy complex user criteria under the graphical modelling framework. We carried out a user study with a web based personal news filtering system, and collected extensive user information, including explicit user feedback, implicit user feedback and some contextual information. Experimental results on the data set collected demonstrate that the graphical modelling approach helps us to better understand the complex domain. The results also show that the complex data driven user modelling approach can improve the adaptive information filtering performance. We also discuss some practical issues while learning complex user models, including how to handle data noise and the missing data problem.
  18. Kang, I.-S.; Na, S.-H.; Lee, S.; Jung, H.; Kim, P.; Sung, W.-K.; Lee, J.-H.: On co-authorship for author disambiguation (2009) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 3453) [ClassicSimilarity], result of:
          0.18536724 = score(doc=3453,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 3453, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=3453)
      0.25 = coord(1/4)
    
    Abstract
    Author name disambiguation deals with clustering the same-name authors into different individuals. To attack the problem, many studies have employed a variety of disambiguation features such as coauthors, titles of papers/publications, topics of articles, emails/affiliations, etc. Among these, co-authorship is the most easily accessible and influential, since inter-person acquaintances represented by co-authorship could discriminate the identities of authors more clearly than other features. This study attempts to explore the net effects of co-authorship on author clustering in bibliographic data. First, to handle the shortage of explicit coauthors listed in known citations, a web-assisted technique of acquiring implicit coauthors of the target author to be disambiguated is proposed. Then, a coauthor disambiguation hypothesis that the identity of an author can be determined by his/her coauthors is examined and confirmed through a variety of author disambiguation experiments.
  19. Stamatatos, E.: ¬A survey of modern authorship attribution methods (2009) 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 3741) [ClassicSimilarity], result of:
          0.18536724 = score(doc=3741,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 3741, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=3741)
      0.25 = coord(1/4)
    
    Abstract
    Authorship attribution supported by statistical or computational methods has a long history starting from the 19th century and is marked by the seminal study of Mosteller and Wallace (1964) on the authorship of the disputed Federalist Papers. During the last decade, this scientific field has been developed substantially, taking advantage of research advances in areas such as machine learning, information retrieval, and natural language processing. The plethora of available electronic texts (e.g., e-mail messages, online forum messages, blogs, source code, etc.) indicates a wide variety of applications of this technology, provided it is able to handle short and noisy text from multiple candidate authors. In this article, a survey of recent advances of the automated approaches to attributing authorship is presented, examining their characteristics for both text representation and text classification. The focus of this survey is on computational requirements and settings rather than on linguistic or literary issues. We also discuss evaluation methodologies and criteria for authorship attribution studies and list open questions that will attract future work in this area.
  20. Correa, C.A.; Kobashi, N.Y.: ¬A hybrid model of automatic indexing based on paraconsitent logic 0.05
    0.04634181 = product of:
      0.18536724 = sum of:
        0.18536724 = weight(_text_:handle in 524) [ClassicSimilarity], result of:
          0.18536724 = score(doc=524,freq=2.0), product of:
            0.42740422 = queryWeight, product of:
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.06532823 = queryNorm
            0.43370473 = fieldWeight in 524, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              6.5424123 = idf(docFreq=173, maxDocs=44421)
              0.046875 = fieldNorm(doc=524)
      0.25 = coord(1/4)
    
    Abstract
    In the processes of information organization, information retrieval and information visualization one point in common can be found: they are strongly connected by the procedures associated to the indexing of texts or documents. Indexing is an essential component of text analysis, and the indexing process have equal importance both for retrieval and for the visualization of information. In this context, it is worth mentioning the solutions that use automatic indexing. The researches that proposes solutions for automatic indexing are based on different theoretical assumptions such as: statistics, linguistics and controlled vocabulary (Leiva 1999). Most solutions developed hybrid models combining these theoretical assumptions. Other solutions to the problem of automatic indexing are based on theories that allow the treatment of uncertainty, imprecision and vagueness. The aim of this paper is to argue the theoretical potential for use in hybrid models of automatic indexing, the paraconsistent logic, a non-classical logic, with capacity to handle situations that involve uncertainty, imprecision and vagueness.

Authors

Languages

Types

  • a 9476
  • m 2245
  • el 1025
  • x 595
  • s 558
  • i 168
  • r 117
  • ? 66
  • n 55
  • b 47
  • l 23
  • p 23
  • h 17
  • d 15
  • u 14
  • fi 10
  • v 2
  • z 2
  • au 1
  • ms 1
  • More… Less…

Themes

Subjects

Classifications