Search (1250 results, page 4 of 63)

  • × language_ss:"e"
  1. Calishain, T.; Dornfest, R.: Google hacks : 100 industrial-strength tips and tools (2003) 0.06
    0.060306095 = product of:
      0.12061219 = sum of:
        0.08638841 = weight(_text_:java in 134) [ClassicSimilarity], result of:
          0.08638841 = score(doc=134,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.19466174 = fieldWeight in 134, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.01953125 = fieldNorm(doc=134)
        0.034223776 = weight(_text_:und in 134) [ClassicSimilarity], result of:
          0.034223776 = score(doc=134,freq=32.0), product of:
            0.13966292 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.06297082 = queryNorm
            0.24504554 = fieldWeight in 134, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.01953125 = fieldNorm(doc=134)
      0.5 = coord(2/4)
    
    Footnote
    Rez. in: nfd - Information Wissenschaft und Praxis 54(2003) H.4, S.253 (D. Lewandowski): "Mit "Google Hacks" liegt das bisher umfassendste Werk vor, das sich ausschließlich an den fortgeschrittenen Google-Nutzer wendet. Daher wird man in diesem Buch auch nicht die sonst üblichen Anfänger-Tips finden, die Suchmaschinenbücher und sonstige Anleitungen zur Internet-Recherche für den professionellen Nutzer in der Regel uninteressant machen. Mit Tara Calishain hat sich eine Autorin gefunden, die bereits seit nahezu fünf Jahren einen eigenen Suchmaschinen-Newsletter (www.researchbuzz.com) herausgibt und als Autorin bzw. Co-Autorin einige Bücher zum Thema Recherche verfasst hat. Für die Programmbeispiele im Buch ist Rael Dornfest verantwortlich. Das erste Kapitel ("Searching Google") gibt einen Einblick in erweiterte Suchmöglichkeiten und Spezifika der behandelten Suchmaschine. Dabei wird der Rechercheansatz der Autorin klar: die beste Methode sei es, die Zahl der Treffer selbst so weit einzuschränken, dass eine überschaubare Menge übrig bleibt, die dann tatsächlich gesichtet werden kann. Dazu werden die feldspezifischen Suchmöglichkeiten in Google erläutert, Tips für spezielle Suchen (nach Zeitschriftenarchiven, technischen Definitionen, usw.) gegeben und spezielle Funktionen der Google-Toolbar erklärt. Bei der Lektüre fällt positiv auf, dass auch der erfahrene Google-Nutzer noch Neues erfährt. Einziges Manko in diesem Kapitel ist der fehlende Blick über den Tellerrand: zwar ist es beispielsweise möglich, mit Google eine Datumssuche genauer als durch das in der erweiterten Suche vorgegebene Auswahlfeld einzuschränken; die aufgezeigte Lösung ist jedoch ausgesprochen umständlich und im Recherchealltag nur eingeschränkt zu gebrauchen. Hier fehlt der Hinweis, dass andere Suchmaschinen weit komfortablere Möglichkeiten der Einschränkung bieten. Natürlich handelt es sich bei dem vorliegenden Werk um ein Buch ausschließlich über Google, trotzdem wäre hier auch ein Hinweis auf die Schwächen hilfreich gewesen. In späteren Kapiteln werden durchaus auch alternative Suchmaschinen zur Lösung einzelner Probleme erwähnt. Das zweite Kapitel widmet sich den von Google neben der klassischen Websuche angebotenen Datenbeständen. Dies sind die Verzeichniseinträge, Newsgroups, Bilder, die Nachrichtensuche und die (hierzulande) weniger bekannten Bereichen Catalogs (Suche in gedruckten Versandhauskatalogen), Froogle (eine in diesem Jahr gestartete Shopping-Suchmaschine) und den Google Labs (hier werden von Google entwickelte neue Funktionen zum öffentlichen Test freigegeben). Nachdem die ersten beiden Kapitel sich ausführlich den Angeboten von Google selbst gewidmet haben, beschäftigt sich das Buch ab Kapitel drei mit den Möglichkeiten, die Datenbestände von Google mittels Programmierungen für eigene Zwecke zu nutzen. Dabei werden einerseits bereits im Web vorhandene Programme vorgestellt, andererseits enthält das Buch viele Listings mit Erläuterungen, um eigene Applikationen zu programmieren. Die Schnittstelle zwischen Nutzer und der Google-Datenbank ist das Google-API ("Application Programming Interface"), das es den registrierten Benutzern erlaubt, täglich bis zu 1.00o Anfragen über ein eigenes Suchinterface an Google zu schicken. Die Ergebnisse werden so zurückgegeben, dass sie maschinell weiterverarbeitbar sind. Außerdem kann die Datenbank in umfangreicherer Weise abgefragt werden als bei einem Zugang über die Google-Suchmaske. Da Google im Gegensatz zu anderen Suchmaschinen in seinen Benutzungsbedingungen die maschinelle Abfrage der Datenbank verbietet, ist das API der einzige Weg, eigene Anwendungen auf Google-Basis zu erstellen. Ein eigenes Kapitel beschreibt die Möglichkeiten, das API mittels unterschiedlicher Programmiersprachen wie PHP, Java, Python, usw. zu nutzen. Die Beispiele im Buch sind allerdings alle in Perl geschrieben, so dass es sinnvoll erscheint, für eigene Versuche selbst auch erst einmal in dieser Sprache zu arbeiten.
    Das sechste Kapitel enthält 26 Anwendungen des Google-APIs, die teilweise von den Autoren des Buchs selbst entwickelt wurden, teils von anderen Autoren ins Netz gestellt wurden. Als besonders nützliche Anwendungen werden unter anderem der Touchgraph Google Browser zur Visualisierung der Treffer und eine Anwendung, die eine Google-Suche mit Abstandsoperatoren erlaubt, vorgestellt. Auffällig ist hier, dass die interessanteren dieser Applikationen nicht von den Autoren des Buchs programmiert wurden. Diese haben sich eher auf einfachere Anwendungen wie beispielsweise eine Zählung der Treffer nach der Top-Level-Domain beschränkt. Nichtsdestotrotz sind auch diese Anwendungen zum großen Teil nützlich. In einem weiteren Kapitel werden pranks and games ("Streiche und Spiele") vorgestellt, die mit dem Google-API realisiert wurden. Deren Nutzen ist natürlich fragwürdig, der Vollständigkeit halber mögen sie in das Buch gehören. Interessanter wiederum ist das letzte Kapitel: "The Webmaster Side of Google". Hier wird Seitenbetreibern erklärt, wie Google arbeitet, wie man Anzeigen am besten formuliert und schaltet, welche Regeln man beachten sollte, wenn man seine Seiten bei Google plazieren will und letztlich auch, wie man Seiten wieder aus dem Google-Index entfernen kann. Diese Ausführungen sind sehr knapp gehalten und ersetzen daher keine Werke, die sich eingehend mit dem Thema Suchmaschinen-Marketing beschäftigen. Allerdings sind die Ausführungen im Gegensatz zu manch anderen Büchern zum Thema ausgesprochen seriös und versprechen keine Wunder in Bezug auf eine Plazienung der eigenen Seiten im Google-Index. "Google Hacks" ist auch denjenigen zu empfehlen, die sich nicht mit der Programmierung mittels des APIs beschäftigen möchten. Dadurch, dass es die bisher umfangreichste Sammlung von Tips und Techniken für einen gezielteren Umgang mit Google darstellt, ist es für jeden fortgeschrittenen Google-Nutzer geeignet. Zwar mögen einige der Hacks einfach deshalb mit aufgenommen worden sein, damit insgesamt die Zahl von i00 erreicht wird. Andere Tips bringen dafür klar erweiterte Möglichkeiten bei der Recherche. Insofern hilft das Buch auch dabei, die für professionelle Bedürfnisse leider unzureichende Abfragesprache von Google ein wenig auszugleichen." - Bergische Landeszeitung Nr.207 vom 6.9.2003, S.RAS04A/1 (Rundschau am Sonntag: Netzwelt) von P. Zschunke: Richtig googeln (s. dort)
  2. Bador, P.; Rey, J.: Description of a professional activity : Modelling of the activity with the completion of a pharmacy thesis related to its terminology environment (2000) 0.05
    0.0527676 = product of:
      0.2110704 = sum of:
        0.2110704 = weight(_text_:handled in 47) [ClassicSimilarity], result of:
          0.2110704 = score(doc=47,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.43031007 = fieldWeight in 47, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.0390625 = fieldNorm(doc=47)
      0.25 = coord(1/4)
    
    Abstract
    In this paper, made tip of parts 1 and 11, we investigate the complex relationships between knowledge, information and activity in order to study how a system of activity assistance can help the actor to solve his information problems. Through the example of the completion of a pharmacy thesis, we have tried, in part 1, to describe, schematize and model the successive phases that make up the whole of this activity. Our method of observation and analysis combined the observation of two students preparing their pharmacy theses, the reading of five theses and the reading of six books. We thus propose in a table form, a modelling outline that presents the sequential succession of the ten operational phases describing the completion of a pharmacy thesis following a chronological order: (1) Subject definition, (2) Documentary research, (3) Documents analysis, (4) Conceiving of the experimental strategy, (5) Experimentation, (6) Results interpretation, (7) Writing of the thesis, (8) Administrative procedures, (9) Preparation of the viva, (10) Viva. The table also presents the succession of the structural, operational, material and human elements: Referents of the activity, Subject of the activity, Location of the activity, Identification of the operations, Handled objects, and Actors. We have refined the activity analysis by drawing up a structured list, showing the organization of the terms related to the different operational phases. This work is presented in part II
  3. Bador, P.; Rey, J.: Description of a professional activity : Modelling of the activity with the completion of a pharmacy thesis related to its terminology environment (2000) 0.05
    0.0527676 = product of:
      0.2110704 = sum of:
        0.2110704 = weight(_text_:handled in 48) [ClassicSimilarity], result of:
          0.2110704 = score(doc=48,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.43031007 = fieldWeight in 48, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.0390625 = fieldNorm(doc=48)
      0.25 = coord(1/4)
    
    Abstract
    Following the modelling of the activity related to the completion of a pharmacy thesis, as presented in part 1, we have completed the activity analysis by drawing up a structured word list to show the terminological organization around the ten operational phases. Indeed, it seemed obvious to us that the reasoning at the root of any activity is based on words used to describe it. This terminology inventory which we called Structured Terminology Environment (STE), together with the modelling diagram, could eventually be directly used during the conceiving of a software tool specific to the studied professional activity. The STE is a thesaurus of 565 words selected on the basis of a corpus stemming from five pharmacy theses and six books, where we put the significant terms which represent the actors and actions we observed during the ten phases, as well as the handled tools. Once the terms were shared out among the ten basic operations, we structured the terminology by grouping the concepts of a same nature so that the sub-categories show a certain homogeneity around the action. We used the following basic relationships: generic/specific relationships, whole/part relationships and finally, we completed the categorization with the help of classes induced by facets (process, phenomenon, properties, material or object, tool or equipment and operating conditions.)
  4. Wijnhoven. F.; Wognum, P.M.; Weg, R.L.W. van de: Knowledge ontology development (1996) 0.05
    0.0527676 = product of:
      0.2110704 = sum of:
        0.2110704 = weight(_text_:handled in 1907) [ClassicSimilarity], result of:
          0.2110704 = score(doc=1907,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.43031007 = fieldWeight in 1907, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1907)
      0.25 = coord(1/4)
    
    Abstract
    Knowledge-containing documents and data about knowledge have been handled in stable environments by bureaucratic systems using very stable knowledge ontologies. These systems, though not always very effective in such environments, will become highly ineffective in environments where knowledge has to be updated and replaced frequently. Moreover, organizations in such dynamic environments also use knowledge from extemal resources extensively. This makes the development of a stable ontology for knowledge storage and retrieval particularly complicated. This paper describes eight context classes of knowledge ontology development and explores elements of a method for ontology development. These classes are based an the differences in contexts defined along three dimensions: knowledge dynamics, complexity and social dispersion. Ontology development matches these contexts and ontology needs defined by (logical and social) structure and ontology maturity. The classification framework and methodology are applied to two cases. The first case illustrates a descriptive use of our framework to characterize ontology development in an academic environment. The second case illustrates a normative use of our framework. The method proposed seemed to be empirically valid and rich and be useful for detecting options for ontology improvement.
  5. Joint, N.: Bemused by bibliometrics : using citation analysis to evaluate research quality (2008) 0.05
    0.0527676 = product of:
      0.2110704 = sum of:
        0.2110704 = weight(_text_:handled in 2900) [ClassicSimilarity], result of:
          0.2110704 = score(doc=2900,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.43031007 = fieldWeight in 2900, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.0390625 = fieldNorm(doc=2900)
      0.25 = coord(1/4)
    
    Abstract
    Purpose - The purpose of this paper is to examine the way in which library and information science (LIS) issues have been handled in the formulation of recent UK Higher Education policy concerned with research quality evaluation. Design/methodology/approach - A chronological review of decision making about digital rights arrangements for the 2008 Research Assessment Exercise (RAE), and of recent announcements about the new shape of metrics-based assessment in the Research Excellence Framework, which supersedes the RAE. Against this chronological framework, the likely nature of LIS practitioner reactions to the flow of decision making is suggested. Findings - It was found that a weak grasp of LIS issues by decision makers undermines the process whereby effective research evaluation models are created. LIS professional opinion should be sampled before key decisions are made. Research limitations/implications - This paper makes no sophisticated comments on the complex research issues underlying advanced bibliometric research evaluation models. It does point out that sophisticated and expensive bibliometric consultancies arrive at many conclusions about metrics-based research assessment that are common knowledge amongst LIS practitioners. Practical implications - Practical difficulties arise when one announces a decision to move to a new and specific type of research evaluation indicator before one has worked out anything very specific about that indicator. Originality/value - In this paper, the importance of information management issues to the mainstream issues of government and public administration is underlined. The most valuable conclusion of this paper is that, because LIS issues are now at the heart of democratic decision making, LIS practitioners and professionals should be given some sort of role in advising on such matters.
  6. Dirks, L.: eResearch, semantic computing and the cloud : towards a smart cyberinfrastructure for eResearch (2009) 0.05
    0.0527676 = product of:
      0.2110704 = sum of:
        0.2110704 = weight(_text_:handled in 3815) [ClassicSimilarity], result of:
          0.2110704 = score(doc=3815,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.43031007 = fieldWeight in 3815, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.0390625 = fieldNorm(doc=3815)
      0.25 = coord(1/4)
    
    Abstract
    In the future, frontier research in many fields will increasingly require the collaboration of globally distributed groups of researchers needing access to distributed computing, data resources and support for remote access to expensive, multi-national specialized facilities such as telescopes and accelerators or specialist data archives. There is also a general belief that an important road to innovation will be provided by multi-disciplinary and collaborative research - from bio-informatics and earth systems science to social science and archaeology. There will also be an explosion in the amount of research data collected in the next decade - 100's of Terabytes will be common in many fields. These future research requirements constitute the 'eResearch' agenda. Powerful software services will be widely deployed on top of the academic research networks to form the necessary 'Cyberinfrastructure' to provide a collaborative research environment for the global academic community. The difficulties in combining data and information from distributed sources, the multi-disciplinary nature of research and collaboration, and the need to move to present researchers with tooling that enable them to express what they want to do rather than how to do it highlight the need for an ecosystem of Semantic Computing technologies. Such technologies will further facilitate information sharing and discovery, will enable reasoning over information, and will allow us to start thinking about knowledge and how it can be handled by computers. This talk will review the elements of this vision and explain the need for semantic-oriented computing by exploring eResearch projects that have successfully applied relevant technologies. It will also suggest that a software + service model with scientific services delivered from the cloud will become an increasingly accepted model for research.
  7. Dervin, B.; Reinhard, C.L.D.: Communication and communication studies (2009) 0.05
    0.0527676 = product of:
      0.2110704 = sum of:
        0.2110704 = weight(_text_:handled in 754) [ClassicSimilarity], result of:
          0.2110704 = score(doc=754,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.43031007 = fieldWeight in 754, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.0390625 = fieldNorm(doc=754)
      0.25 = coord(1/4)
    
    Abstract
    The focal purpose of this entry is to lay out the contours, within historical contexts, of the structural arrangements and intellectual foci of the various areas of study and emphases in institutions of higher education generally clustered under the much used and abused term "communication." Because communication studies has been a presence in the academy for less than 60 years, and because it has evolved in the midst of an exploding and rapidly changing communication environment, there are necessarily multiple and competing stories to be told. This entry attempts to focus both on convergences and divergences. The entry reviews in turn: different meanings for the word "communication" and how and when the term came into popular usage; the two primary historical lineages out of which communication studies has emerged-speech communication and mass communication; the forces leading to institutionalization, including the exploding emergence of mass media. The entry examines the phases in which communication studies became institutionalized and how supporting structures of organizations and journals have handled seemingly overwhelming diversity. The entry concludes with a discussion of how communication studies necessarily holds within its core a series of contradictions that are in essence symptom and cause of both centralities and dispersions, struggles and successes.
  8. Noy, N.F.; Musen, M.A.: PROMPT: algorithm and tool for automated ontology merging and alignment 0.05
    0.0527676 = product of:
      0.2110704 = sum of:
        0.2110704 = weight(_text_:handled in 900) [ClassicSimilarity], result of:
          0.2110704 = score(doc=900,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.43031007 = fieldWeight in 900, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.0390625 = fieldNorm(doc=900)
      0.25 = coord(1/4)
    
    Abstract
    Researchers in the ontology-design field have developed the content for ontologies in many domain areas. Recently, ontologies have become increasingly common on the World- Wide Web where they provide semantics for annotations in Web pages. This distributed nature of ontology development has led to a large number of ontologies covering overlapping domains. In order for these ontologies to be reused, they first need to be merged or aligned to one another. The processes of ontology alignment and merging are usually handled manually and often constitute a large and tedious portion of the sharing process. We have developed and implemented PROMPT, an algorithm that provides a semi-automatic approach to ontology merging and alignment. PROMPT performs some tasks automatically and guides the user in performing other tasks for which his intervention is required. PROMPT also determines possible inconsistencies in the state of the ontology, which result from the user's actions, and suggests ways to remedy these inconsistencies. PROMPT is based on an extremely general knowledge model and therefore can be applied across various platforms. Our formative evaluation showed that a human expert followed 90% of the suggestions that PROMPT generated and that 74% of the total knowledge-base operations invoked by the user were suggested by PROMPT.
  9. Broughton, V.: Concepts and terms in the faceted classification : the case of UDC (2010) 0.05
    0.0527676 = product of:
      0.2110704 = sum of:
        0.2110704 = weight(_text_:handled in 65) [ClassicSimilarity], result of:
          0.2110704 = score(doc=65,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.43031007 = fieldWeight in 65, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.0390625 = fieldNorm(doc=65)
      0.25 = coord(1/4)
    
    Abstract
    Recent revision of UDC classes has aimed at implementing a more faceted approach. Many compound classes have been removed from the main tables, and more radical revisions of classes (particularly those for Medicine and Religion) have introduced a rigorous analysis, a clearer sense of citation order, and building of compound classes according to a more logical system syntax. The faceted approach provides a means of formalizing the relationships in the classification and making them explicit for machine recognition. In the Bliss Bibliographic Classification (BC2) (which has been a source for both UDC classes mentioned above), terminologies are encoded for automatic generation of hierarchical and associative relationships. Nevertheless, difficulties are encountered in vocabulary control, and a similar phenomenon is observed in UDC. Current work has revealed differences in the vocabulary of humanities and science, notably the way in which terms in the humanities should be handled when these are semantically complex. Achieving a balance between rigour in the structure of the classification and the complexity of natural language expression remains partially unresolved at present, but provides a fertile field for further research.
  10. Bounhas, I.; Elayeb, B.; Evrard, F.; Slimani, Y.: Organizing contextual knowledge for Arabic text disambiguation and terminology extraction (2011) 0.05
    0.0527676 = product of:
      0.2110704 = sum of:
        0.2110704 = weight(_text_:handled in 846) [ClassicSimilarity], result of:
          0.2110704 = score(doc=846,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.43031007 = fieldWeight in 846, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.0390625 = fieldNorm(doc=846)
      0.25 = coord(1/4)
    
    Abstract
    Ontologies have an important role in knowledge organization and information retrieval. Domain ontologies are composed of concepts represented by domain relevant terms. Existing approaches of ontology construction make use of statistical and linguistic information to extract domain relevant terms. The quality and the quantity of this information influence the accuracy of terminology extraction approaches and other steps in knowledge extraction and information retrieval. This paper proposes an approach for handling domain relevant terms from Arabic non-diacriticised semi-structured corpora. In input, the structure of documents is exploited to organize knowledge in a contextual graph, which is exploited to extract relevant terms. This network contains simple and compound nouns handled by a morphosyntactic shallow parser. The noun phrases are evaluated in terms of termhood and unithood by means of possibilistic measures. We apply a qualitative approach, which weighs terms according to their positions in the structure of the document. In output, the extracted knowledge is organized as network modeling dependencies between terms, which can be exploited to infer semantic relations. We test our approach on three specific domain corpora. The goal of this evaluation is to check if our model for organizing and exploiting contextual knowledge will improve the accuracy of extraction of simple and compound nouns. We also investigate the role of compound nouns in improving information retrieval results.
  11. Borgman, C.L.: ¬The conundrum of sharing research data (2012) 0.05
    0.0527676 = product of:
      0.2110704 = sum of:
        0.2110704 = weight(_text_:handled in 1248) [ClassicSimilarity], result of:
          0.2110704 = score(doc=1248,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.43031007 = fieldWeight in 1248, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1248)
      0.25 = coord(1/4)
    
    Abstract
    Researchers are producing an unprecedented deluge of data by using new methods and instrumentation. Others may wish to mine these data for new discoveries and innovations. However, research data are not readily available as sharing is common in only a few fields such as astronomy and genomics. Data sharing practices in other fields vary widely. Moreover, research data take many forms, are handled in many ways, using many approaches, and often are difficult to interpret once removed from their initial context. Data sharing is thus a conundrum. Four rationales for sharing data are examined, drawing examples from the sciences, social sciences, and humanities: (1) to reproduce or to verify research, (2) to make results of publicly funded research available to the public, (3) to enable others to ask new questions of extant data, and (4) to advance the state of research and innovation. These rationales differ by the arguments for sharing, by beneficiaries, and by the motivations and incentives of the many stakeholders involved. The challenges are to understand which data might be shared, by whom, with whom, under what conditions, why, and to what effects. Answers will inform data policy and practice.
  12. Austin, D.: Prospects for a new general classification (1969) 0.05
    0.0527676 = product of:
      0.2110704 = sum of:
        0.2110704 = weight(_text_:handled in 2519) [ClassicSimilarity], result of:
          0.2110704 = score(doc=2519,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.43031007 = fieldWeight in 2519, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.0390625 = fieldNorm(doc=2519)
      0.25 = coord(1/4)
    
    Abstract
    In traditional classification schemes, the universe of knowledge is brokeii down into self- contained disciplines which are further analysed to the point at which a particular concept is located. This leads to problems of: (a) currency: keeping the scheme in line with new discoveries. (b) hospitality: allowing room for insertion of new subjects (c) cross-classification: a concept may be considered in such a way that it fits as logically into one discipline as another. Machine retrieval is also hampered by the fact that any individual concept is notated differently, depending on where in the scheme it appears. The approach now considered is from an organized universe of concepts, every concept being set down only once in an appropriate vocabulary, where it acquires the notation which identifies it wherever it is used. It has been found that all the concepts present in any compound subject can be handled as though they belong to one of two basic concept types, being either Entities or Attributes. In classing, these concepts are identified, and notation is selected from appropriate schedules. Subjects are then built according to formal rules, the final class number incorporating operators which convey the fundamental relationships between concepts. From this viewpoint, the Rules and Operators of the proposed system can be seen as the grammar of an IR language, and the schedules of Entities and Attributes as its vocabulary.
  13. Greyson, D.L.; Johnson, J.L.: ¬The role of information in health behavior : a scoping study and discussion of major public health models (2016) 0.05
    0.0527676 = product of:
      0.2110704 = sum of:
        0.2110704 = weight(_text_:handled in 4218) [ClassicSimilarity], result of:
          0.2110704 = score(doc=4218,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.43031007 = fieldWeight in 4218, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.0390625 = fieldNorm(doc=4218)
      0.25 = coord(1/4)
    
    Abstract
    Information interventions that influence health behavior are a major element of the public health toolkit and an area of potential interest and investigation for library and information science (LIS) researchers. To explore the use of information as a concept within dominant public health behavior models and the manner in which information practices are handled therein, we undertook a scoping study. We scoped the use of "information" within core English-language health behavior textbooks and examined dominant models of health behavior for information practices. Index terms within these texts indicated a lack of common language around information-related concepts. Nine models/theories were discussed in a majority of the texts. These were grouped by model type and examined for information-related concepts/constructs. Information was framed as a "thing" or resource, and information practices were commonly included or implied. However, lack of specificity regarding the definition of information, how it differs from knowledge, and how context affects information practices make the exact role of information within health behavior models unclear. Although health information interventions may be grounded in behavioral theory, a limited understanding of the ways information works within people's lives hinders our ability to effectively use information to improve health. By the same token, information scientists should explore public health's interventionist approach.
  14. Hasanain, M.; Elsayed, T.: Studying effectiveness of Web search for fact checking (2022) 0.05
    0.0527676 = product of:
      0.2110704 = sum of:
        0.2110704 = weight(_text_:handled in 1559) [ClassicSimilarity], result of:
          0.2110704 = score(doc=1559,freq=2.0), product of:
            0.4905077 = queryWeight, product of:
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.06297082 = queryNorm
            0.43031007 = fieldWeight in 1559, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.7894444 = idf(docFreq=49, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1559)
      0.25 = coord(1/4)
    
    Abstract
    Web search is commonly used by fact checking systems as a source of evidence for claim verification. In this work, we demonstrate that the task of retrieving pages useful for fact checking, called evidential pages, is indeed different from the task of retrieving topically relevant pages that are typically optimized by search engines; thus, it should be handled differently. We conduct a comprehensive study on the performance of retrieving evidential pages over a test collection we developed for the task of re-ranking Web pages by usefulness for fact-checking. Results show that pages (retrieved by a commercial search engine) that are topically relevant to a claim are not always useful for verifying it, and that the engine's performance in retrieving evidential pages is weakly correlated with retrieval of topically relevant pages. Additionally, we identify types of evidence in evidential pages and some linguistic cues that can help predict page usefulness. Moreover, preliminary experiments show that a retrieval model leveraging those cues has a higher performance compared to the search engine. Finally, we show that existing systems have a long way to go to support effective fact checking. To that end, our work provides insights to guide design of better future systems for the task.
  15. Gibson, P.: Professionals' perfect Web world in sight : users want more information on the Web, and vendors attempt to provide (1998) 0.05
    0.05183304 = product of:
      0.20733216 = sum of:
        0.20733216 = weight(_text_:java in 2656) [ClassicSimilarity], result of:
          0.20733216 = score(doc=2656,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.46718815 = fieldWeight in 2656, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=2656)
      0.25 = coord(1/4)
    
    Abstract
    Many information professionals feel that the time is still far off when the WWW can offer the combined funtionality and content of traditional online and CD-ROM databases, but there have been a number of recent Web developments to reflect on. Describes the testing and launch by Ovid of its Java client which, in effect, allows access to its databases on the Web with full search functionality, and the initiative of Euromonitor in providing Web access to its whole collection of consumer research reports and its entire database of business sources. Also reviews the service of a newcomer to the information scene, Information Quest (IQ) founded by Dawson Holdings which has made an agreement with Infonautics to offer access to its Electric Library database thus adding over 1.000 reference, consumer and business publications to its Web based journal service
  16. Nieuwenhuysen, P.; Vanouplines, P.: Document plus program hybrids on the Internet and their impact on information transfer (1998) 0.05
    0.05183304 = product of:
      0.20733216 = sum of:
        0.20733216 = weight(_text_:java in 2893) [ClassicSimilarity], result of:
          0.20733216 = score(doc=2893,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.46718815 = fieldWeight in 2893, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=2893)
      0.25 = coord(1/4)
    
    Abstract
    Examines some of the advanced tools, techniques, methods and standards related to the Internet and WWW which consist of hybrids of documents and software, called 'document program hybrids'. Early Internet systems were based on having documents on one side and software on the other, neatly separated, apart from one another and without much interaction, so that the static document can also exist without computers and networks. Documentation program hybrids blur this classical distinction and all components are integrated, interwoven and exist in synergy with each other. Illustrates the techniques with particular reference to practical examples, including: dara collections and dedicated software; advanced HTML features on the WWW, multimedia viewer and plug in software for Internet and WWW browsers; VRML; interaction through a Web server with other servers and with instruments; adaptive hypertext provided by the server; 'webbots' or 'knowbots' or 'searchbots' or 'metasearch engines' or intelligent software agents; Sun's Java; Microsoft's ActiveX; program scripts for HTML and Web browsers; cookies; and Internet push technology with Webcasting channels
  17. Mills, T.; Moody, K.; Rodden, K.: Providing world wide access to historical sources (1997) 0.05
    0.05183304 = product of:
      0.20733216 = sum of:
        0.20733216 = weight(_text_:java in 3697) [ClassicSimilarity], result of:
          0.20733216 = score(doc=3697,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.46718815 = fieldWeight in 3697, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=3697)
      0.25 = coord(1/4)
    
    Abstract
    A unique collection of historical material covering the lives and events of an English village between 1400 and 1750 has been made available via a WWW enabled information retrieval system. Since the expected readership of the documents ranges from school children to experienced researchers, providing this information in an easily accessible form has offered many challenges requiring tools to aid searching and browsing. The file structure of the document collection was replaced by an database, enabling query results to be presented on the fly. A Java interface displays each user's context in a form that allows for easy and intuitive relevance feedback
  18. Maarek, Y.S.: WebCutter : a system for dynamic and tailorable site mapping (1997) 0.05
    0.05183304 = product of:
      0.20733216 = sum of:
        0.20733216 = weight(_text_:java in 3739) [ClassicSimilarity], result of:
          0.20733216 = score(doc=3739,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.46718815 = fieldWeight in 3739, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=3739)
      0.25 = coord(1/4)
    
    Abstract
    Presents an approach that integrates searching and browsing in a manner that improves both paradigms. When browsing is the primary task, it enables semantic content-based tailoring of Web maps in both the generation as well as the visualization phases. When search is the primary task, it enables contextualization of the results by augmenting them with the documents' neighbourhoods. This approach is embodied in WebCutter, a client-server system fully integrated with Web software. WebCutter consists of a map generator running off a standard Web server and a map visualization client implemented as a Java applet runalble from any standard Web browser and requiring no installation or external plug-in application. WebCutter is in beta stage and is in the process of being integrated into the Lotus Domino application product line
  19. Pan, B.; Gay, G.; Saylor, J.; Hembrooke, H.: One digital library, two undergraduate casses, and four learning modules : uses of a digital library in cassrooms (2006) 0.05
    0.05183304 = product of:
      0.20733216 = sum of:
        0.20733216 = weight(_text_:java in 907) [ClassicSimilarity], result of:
          0.20733216 = score(doc=907,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.46718815 = fieldWeight in 907, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=907)
      0.25 = coord(1/4)
    
    Abstract
    The KMODDL (kinematic models for design digital library) is a digital library based on a historical collection of kinematic models made of steel and bronze. The digital library contains four types of learning modules including textual materials, QuickTime virtual reality movies, Java simulations, and stereolithographic files of the physical models. The authors report an evaluation study on the uses of the KMODDL in two undergraduate classes. This research reveals that the users in different classes encountered different usability problems, and reported quantitatively different subjective experiences. Further, the results indicate that depending on the subject area, the two user groups preferred different types of learning modules, resulting in different uses of the available materials and different learning outcomes. These findings are discussed in terms of their implications for future digital library design.
  20. Mongin, L.; Fu, Y.Y.; Mostafa, J.: Open Archives data Service prototype and automated subject indexing using D-Lib archive content as a testbed (2003) 0.05
    0.05183304 = product of:
      0.20733216 = sum of:
        0.20733216 = weight(_text_:java in 2167) [ClassicSimilarity], result of:
          0.20733216 = score(doc=2167,freq=2.0), product of:
            0.4437873 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06297082 = queryNorm
            0.46718815 = fieldWeight in 2167, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=2167)
      0.25 = coord(1/4)
    
    Abstract
    The Indiana University School of Library and Information Science opened a new research laboratory in January 2003; The Indiana University School of Library and Information Science Information Processing Laboratory [IU IP Lab]. The purpose of the new laboratory is to facilitate collaboration between scientists in the department in the areas of information retrieval (IR) and information visualization (IV) research. The lab has several areas of focus. These include grid and cluster computing, and a standard Java-based software platform to support plug and play research datasets, a selection of standard IR modules and standard IV algorithms. Future development includes software to enable researchers to contribute datasets, IR algorithms, and visualization algorithms into the standard environment. We decided early on to use OAI-PMH as a resource discovery tool because it is consistent with our mission.

Languages

  • d 32
  • m 3
  • nl 1
  • More… Less…

Types

  • a 807
  • m 313
  • el 106
  • s 93
  • i 21
  • n 17
  • x 14
  • r 10
  • b 8
  • ? 1
  • v 1
  • More… Less…

Themes

Subjects

Classifications