Search (13237 results, page 6 of 662)

  1. Hancock, B.; Giarlo, M.J.: Moving to XML : Latin texts XML conversion project at the Center for Electronic Texts in the Humanities (2001) 0.06
    0.059572075 = product of:
      0.2382883 = sum of:
        0.2382883 = weight(_text_:java in 5801) [ClassicSimilarity], result of:
          0.2382883 = score(doc=5801,freq=2.0), product of:
            0.43718386 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06203383 = queryNorm
            0.5450528 = fieldWeight in 5801, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0546875 = fieldNorm(doc=5801)
      0.25 = coord(1/4)
    
    Abstract
    The delivery of documents on the Web has moved beyond the restrictions of the traditional Web markup language, HTML. HTML's static tags cannot deal with the variety of data formats now beginning to be exchanged between various entities, whether corporate or institutional. XML solves many of the problems by allowing arbitrary tags, which describe the content for a particular audience or group. At the Center for Electronic Texts in the Humanities the Latin texts of Lector Longinquus are being transformed to XML in readiness for the expected new standard. To allow existing browsers to render these texts, a Java program is used to transform the XML to HTML on the fly.
  2. Calishain, T.; Dornfest, R.: Google hacks : 100 industrial-strength tips and tools (2003) 0.06
    0.059408754 = product of:
      0.11881751 = sum of:
        0.08510297 = weight(_text_:java in 134) [ClassicSimilarity], result of:
          0.08510297 = score(doc=134,freq=2.0), product of:
            0.43718386 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06203383 = queryNorm
            0.19466174 = fieldWeight in 134, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.01953125 = fieldNorm(doc=134)
        0.033714537 = weight(_text_:und in 134) [ClassicSimilarity], result of:
          0.033714537 = score(doc=134,freq=32.0), product of:
            0.13758478 = queryWeight, product of:
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.06203383 = queryNorm
            0.24504554 = fieldWeight in 134, product of:
              5.656854 = tf(freq=32.0), with freq of:
                32.0 = termFreq=32.0
              2.217899 = idf(docFreq=13141, maxDocs=44421)
              0.01953125 = fieldNorm(doc=134)
      0.5 = coord(2/4)
    
    Footnote
    Rez. in: nfd - Information Wissenschaft und Praxis 54(2003) H.4, S.253 (D. Lewandowski): "Mit "Google Hacks" liegt das bisher umfassendste Werk vor, das sich ausschließlich an den fortgeschrittenen Google-Nutzer wendet. Daher wird man in diesem Buch auch nicht die sonst üblichen Anfänger-Tips finden, die Suchmaschinenbücher und sonstige Anleitungen zur Internet-Recherche für den professionellen Nutzer in der Regel uninteressant machen. Mit Tara Calishain hat sich eine Autorin gefunden, die bereits seit nahezu fünf Jahren einen eigenen Suchmaschinen-Newsletter (www.researchbuzz.com) herausgibt und als Autorin bzw. Co-Autorin einige Bücher zum Thema Recherche verfasst hat. Für die Programmbeispiele im Buch ist Rael Dornfest verantwortlich. Das erste Kapitel ("Searching Google") gibt einen Einblick in erweiterte Suchmöglichkeiten und Spezifika der behandelten Suchmaschine. Dabei wird der Rechercheansatz der Autorin klar: die beste Methode sei es, die Zahl der Treffer selbst so weit einzuschränken, dass eine überschaubare Menge übrig bleibt, die dann tatsächlich gesichtet werden kann. Dazu werden die feldspezifischen Suchmöglichkeiten in Google erläutert, Tips für spezielle Suchen (nach Zeitschriftenarchiven, technischen Definitionen, usw.) gegeben und spezielle Funktionen der Google-Toolbar erklärt. Bei der Lektüre fällt positiv auf, dass auch der erfahrene Google-Nutzer noch Neues erfährt. Einziges Manko in diesem Kapitel ist der fehlende Blick über den Tellerrand: zwar ist es beispielsweise möglich, mit Google eine Datumssuche genauer als durch das in der erweiterten Suche vorgegebene Auswahlfeld einzuschränken; die aufgezeigte Lösung ist jedoch ausgesprochen umständlich und im Recherchealltag nur eingeschränkt zu gebrauchen. Hier fehlt der Hinweis, dass andere Suchmaschinen weit komfortablere Möglichkeiten der Einschränkung bieten. Natürlich handelt es sich bei dem vorliegenden Werk um ein Buch ausschließlich über Google, trotzdem wäre hier auch ein Hinweis auf die Schwächen hilfreich gewesen. In späteren Kapiteln werden durchaus auch alternative Suchmaschinen zur Lösung einzelner Probleme erwähnt. Das zweite Kapitel widmet sich den von Google neben der klassischen Websuche angebotenen Datenbeständen. Dies sind die Verzeichniseinträge, Newsgroups, Bilder, die Nachrichtensuche und die (hierzulande) weniger bekannten Bereichen Catalogs (Suche in gedruckten Versandhauskatalogen), Froogle (eine in diesem Jahr gestartete Shopping-Suchmaschine) und den Google Labs (hier werden von Google entwickelte neue Funktionen zum öffentlichen Test freigegeben). Nachdem die ersten beiden Kapitel sich ausführlich den Angeboten von Google selbst gewidmet haben, beschäftigt sich das Buch ab Kapitel drei mit den Möglichkeiten, die Datenbestände von Google mittels Programmierungen für eigene Zwecke zu nutzen. Dabei werden einerseits bereits im Web vorhandene Programme vorgestellt, andererseits enthält das Buch viele Listings mit Erläuterungen, um eigene Applikationen zu programmieren. Die Schnittstelle zwischen Nutzer und der Google-Datenbank ist das Google-API ("Application Programming Interface"), das es den registrierten Benutzern erlaubt, täglich bis zu 1.00o Anfragen über ein eigenes Suchinterface an Google zu schicken. Die Ergebnisse werden so zurückgegeben, dass sie maschinell weiterverarbeitbar sind. Außerdem kann die Datenbank in umfangreicherer Weise abgefragt werden als bei einem Zugang über die Google-Suchmaske. Da Google im Gegensatz zu anderen Suchmaschinen in seinen Benutzungsbedingungen die maschinelle Abfrage der Datenbank verbietet, ist das API der einzige Weg, eigene Anwendungen auf Google-Basis zu erstellen. Ein eigenes Kapitel beschreibt die Möglichkeiten, das API mittels unterschiedlicher Programmiersprachen wie PHP, Java, Python, usw. zu nutzen. Die Beispiele im Buch sind allerdings alle in Perl geschrieben, so dass es sinnvoll erscheint, für eigene Versuche selbst auch erst einmal in dieser Sprache zu arbeiten.
    Das sechste Kapitel enthält 26 Anwendungen des Google-APIs, die teilweise von den Autoren des Buchs selbst entwickelt wurden, teils von anderen Autoren ins Netz gestellt wurden. Als besonders nützliche Anwendungen werden unter anderem der Touchgraph Google Browser zur Visualisierung der Treffer und eine Anwendung, die eine Google-Suche mit Abstandsoperatoren erlaubt, vorgestellt. Auffällig ist hier, dass die interessanteren dieser Applikationen nicht von den Autoren des Buchs programmiert wurden. Diese haben sich eher auf einfachere Anwendungen wie beispielsweise eine Zählung der Treffer nach der Top-Level-Domain beschränkt. Nichtsdestotrotz sind auch diese Anwendungen zum großen Teil nützlich. In einem weiteren Kapitel werden pranks and games ("Streiche und Spiele") vorgestellt, die mit dem Google-API realisiert wurden. Deren Nutzen ist natürlich fragwürdig, der Vollständigkeit halber mögen sie in das Buch gehören. Interessanter wiederum ist das letzte Kapitel: "The Webmaster Side of Google". Hier wird Seitenbetreibern erklärt, wie Google arbeitet, wie man Anzeigen am besten formuliert und schaltet, welche Regeln man beachten sollte, wenn man seine Seiten bei Google plazieren will und letztlich auch, wie man Seiten wieder aus dem Google-Index entfernen kann. Diese Ausführungen sind sehr knapp gehalten und ersetzen daher keine Werke, die sich eingehend mit dem Thema Suchmaschinen-Marketing beschäftigen. Allerdings sind die Ausführungen im Gegensatz zu manch anderen Büchern zum Thema ausgesprochen seriös und versprechen keine Wunder in Bezug auf eine Plazienung der eigenen Seiten im Google-Index. "Google Hacks" ist auch denjenigen zu empfehlen, die sich nicht mit der Programmierung mittels des APIs beschäftigen möchten. Dadurch, dass es die bisher umfangreichste Sammlung von Tips und Techniken für einen gezielteren Umgang mit Google darstellt, ist es für jeden fortgeschrittenen Google-Nutzer geeignet. Zwar mögen einige der Hacks einfach deshalb mit aufgenommen worden sein, damit insgesamt die Zahl von i00 erreicht wird. Andere Tips bringen dafür klar erweiterte Möglichkeiten bei der Recherche. Insofern hilft das Buch auch dabei, die für professionelle Bedürfnisse leider unzureichende Abfragesprache von Google ein wenig auszugleichen." - Bergische Landeszeitung Nr.207 vom 6.9.2003, S.RAS04A/1 (Rundschau am Sonntag: Netzwelt) von P. Zschunke: Richtig googeln (s. dort)
  3. McIlwaine, I.C.: ¬The Universal Decimal Classification : a guide to its use (2000) 0.06
    0.058558512 = product of:
      0.23423405 = sum of:
        0.23423405 = weight(_text_:hints in 1161) [ClassicSimilarity], result of:
          0.23423405 = score(doc=1161,freq=2.0), product of:
            0.5128635 = queryWeight, product of:
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.06203383 = queryNorm
            0.4567181 = fieldWeight in 1161, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1161)
      0.25 = coord(1/4)
    
    Abstract
    This book is an extension and total revision of the author's earlier Guide to the use of UDC. The original was written in 1993 and in the intervening years much has happened with the classification. In particular, a much more rigorous approach has been undertaken in revision to ensure that the scheme is able to handle the requirements of a networked world. The book outlines the history and development of the Universal Decimal Classification, provides practical hints on its application and works through all the auxiliary and main tables highlighting aspects that need to be noted in applying the scheme. It also provides guidance on the use of the Master Reference File and discusses the ways in which the classification is used in the 21st century and its suitability as an aid to subject description in tagging metadata and consequently for application on the Internet. It is intended as a source for information about the scheme, for practical usage by classifiers in their daily work and as a guide to the student learning how to apply the classification. It is amply provided with examples to illustrate the many ways in which the scheme can be applied and will be a useful source for a wide range of information workers
  4. Müller, J.F.: ¬A librarian's guide to the Internet : a guide to searching and evaluating information (2003) 0.06
    0.058558512 = product of:
      0.23423405 = sum of:
        0.23423405 = weight(_text_:hints in 5502) [ClassicSimilarity], result of:
          0.23423405 = score(doc=5502,freq=2.0), product of:
            0.5128635 = queryWeight, product of:
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.06203383 = queryNorm
            0.4567181 = fieldWeight in 5502, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.0390625 = fieldNorm(doc=5502)
      0.25 = coord(1/4)
    
    Content
    Key Features - Helps a librarian deliver perfect customer service with confidence - Provides practical tips and hints; is pragmatic rather than technical - Is written by a highly respected and experienced practitioner The Author Jeanne Froidevaux Müller is a frequent contributor to the respected magazine Managing Information; the author was, from 1992-2002, head of the library at the Swiss Cancer League. Jeanne is currently based at the public library of Thun, Switzerland. Readership The book is aimed at all librarians and informational professionals: in the academic, public and private sectors. It will be of interest to both large and small libraries. Contents Introduction Basis of confidence - the Internet as a tool and not something to be afraid of Data - Information - Knowledge How to search - simple strategies; what to ask; examples Interpreting results-including examples Maintaining a link list an your browser How to teach your customers what you know and how to know what your customers want Perfect customer service
  5. Porter, K.: Setting up a new library and information service (2003) 0.06
    0.058558512 = product of:
      0.23423405 = sum of:
        0.23423405 = weight(_text_:hints in 5510) [ClassicSimilarity], result of:
          0.23423405 = score(doc=5510,freq=2.0), product of:
            0.5128635 = queryWeight, product of:
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.06203383 = queryNorm
            0.4567181 = fieldWeight in 5510, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.0390625 = fieldNorm(doc=5510)
      0.25 = coord(1/4)
    
    Abstract
    Is intended as a guide for those people who have been charged with establishing a library or information service in their organisation. It is a practical guide to the steps that need to be taken, the people that will need to be dealt with and the procedures that will have to be adopted in order to bring the venture to a successful conclusion. Contains many practical tips and hints from the author.
  6. Dawson, H.: Using the Internet for political research : practical tips and hints (2003) 0.06
    0.058558512 = product of:
      0.23423405 = sum of:
        0.23423405 = weight(_text_:hints in 5511) [ClassicSimilarity], result of:
          0.23423405 = score(doc=5511,freq=2.0), product of:
            0.5128635 = queryWeight, product of:
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.06203383 = queryNorm
            0.4567181 = fieldWeight in 5511, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.0390625 = fieldNorm(doc=5511)
      0.25 = coord(1/4)
    
  7. Warner, J.: What should we understand by information technology (and some hints at other issues)? (2000) 0.06
    0.058558512 = product of:
      0.23423405 = sum of:
        0.23423405 = weight(_text_:hints in 840) [ClassicSimilarity], result of:
          0.23423405 = score(doc=840,freq=2.0), product of:
            0.5128635 = queryWeight, product of:
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.06203383 = queryNorm
            0.4567181 = fieldWeight in 840, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.0390625 = fieldNorm(doc=840)
      0.25 = coord(1/4)
    
  8. Liu, J.S.; Lu, L.Y.Y.: ¬An integrated approach for main path analysis : development of the Hirsch index as an example (2012) 0.06
    0.058558512 = product of:
      0.23423405 = sum of:
        0.23423405 = weight(_text_:hints in 1072) [ClassicSimilarity], result of:
          0.23423405 = score(doc=1072,freq=2.0), product of:
            0.5128635 = queryWeight, product of:
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.06203383 = queryNorm
            0.4567181 = fieldWeight in 1072, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1072)
      0.25 = coord(1/4)
    
    Abstract
    This study enhances main path analysis by proposing several variants to the original approach. Main path analysis is a bibliometric method capable of tracing the most significant paths in a citation network and is commonly used to trace the development trajectory of a research field. We highlight several limitations of the original main path analysis and suggest new, complementary approaches to overcome these limitations. In contrast to the original local main path, the new approaches generate the global main path, the backward local main path, multiple main paths, and key-route main paths. Each of them is obtained via a perspective different from the original approach. By simultaneously conducting the new, complementary approaches, one uncovers the key development of the target discipline from a broader view. To demonstrate the value of these new approaches, we simultaneously apply them to a set of academic articles related to the Hirsch index. The results show that the integrated approach discovers several paths that are not captured by the original approach. Among these new approaches, the key-route approach is especially useful and hints at a divergence-convergence-divergence structure in the development of the Hirsch index.
  9. Vakkari, P.; Chang, Y.-W.; Järvelin, K.: Disciplinary contributions to research topics and methodology in Library and Information Science : leading to fragmentation? (2022) 0.06
    0.058558512 = product of:
      0.23423405 = sum of:
        0.23423405 = weight(_text_:hints in 1768) [ClassicSimilarity], result of:
          0.23423405 = score(doc=1768,freq=2.0), product of:
            0.5128635 = queryWeight, product of:
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.06203383 = queryNorm
            0.4567181 = fieldWeight in 1768, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1768)
      0.25 = coord(1/4)
    
    Abstract
    The study analyses contributions to Library and Information Science (LIS) by researchers representing various disciplines. How are such contributions associated with the choice of research topics and methodology? The study employs a quantitative content analysis of articles published in 31 scholarly LIS journals in 2015. Each article is seen as a contribution to LIS by the authors' disciplines, which are inferred from their affiliations. The unit of analysis is the article-discipline pair. Of the contribution instances, the share of LIS is one third. Computer Science contributes one fifth and Business and Economics one sixth. The latter disciplines dominate the contributions in information retrieval, information seeking, and scientific communication indicating strong influences in LIS. Correspondence analysis reveals three clusters of research, one focusing on traditional LIS with contributions from LIS and Humanities and survey-type research; another on information retrieval with contributions from Computer Science and experimental research; and the third on scientific communication with contributions from Natural Sciences and Medicine and citation analytic research. The strong differentiation of scholarly contributions in LIS hints to the fragmentation of LIS as a discipline.
  10. Gibson, P.: Professionals' perfect Web world in sight : users want more information on the Web, and vendors attempt to provide (1998) 0.05
    0.05106178 = product of:
      0.20424712 = sum of:
        0.20424712 = weight(_text_:java in 2656) [ClassicSimilarity], result of:
          0.20424712 = score(doc=2656,freq=2.0), product of:
            0.43718386 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06203383 = queryNorm
            0.46718815 = fieldWeight in 2656, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=2656)
      0.25 = coord(1/4)
    
    Abstract
    Many information professionals feel that the time is still far off when the WWW can offer the combined funtionality and content of traditional online and CD-ROM databases, but there have been a number of recent Web developments to reflect on. Describes the testing and launch by Ovid of its Java client which, in effect, allows access to its databases on the Web with full search functionality, and the initiative of Euromonitor in providing Web access to its whole collection of consumer research reports and its entire database of business sources. Also reviews the service of a newcomer to the information scene, Information Quest (IQ) founded by Dawson Holdings which has made an agreement with Infonautics to offer access to its Electric Library database thus adding over 1.000 reference, consumer and business publications to its Web based journal service
  11. Nieuwenhuysen, P.; Vanouplines, P.: Document plus program hybrids on the Internet and their impact on information transfer (1998) 0.05
    0.05106178 = product of:
      0.20424712 = sum of:
        0.20424712 = weight(_text_:java in 2893) [ClassicSimilarity], result of:
          0.20424712 = score(doc=2893,freq=2.0), product of:
            0.43718386 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06203383 = queryNorm
            0.46718815 = fieldWeight in 2893, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=2893)
      0.25 = coord(1/4)
    
    Abstract
    Examines some of the advanced tools, techniques, methods and standards related to the Internet and WWW which consist of hybrids of documents and software, called 'document program hybrids'. Early Internet systems were based on having documents on one side and software on the other, neatly separated, apart from one another and without much interaction, so that the static document can also exist without computers and networks. Documentation program hybrids blur this classical distinction and all components are integrated, interwoven and exist in synergy with each other. Illustrates the techniques with particular reference to practical examples, including: dara collections and dedicated software; advanced HTML features on the WWW, multimedia viewer and plug in software for Internet and WWW browsers; VRML; interaction through a Web server with other servers and with instruments; adaptive hypertext provided by the server; 'webbots' or 'knowbots' or 'searchbots' or 'metasearch engines' or intelligent software agents; Sun's Java; Microsoft's ActiveX; program scripts for HTML and Web browsers; cookies; and Internet push technology with Webcasting channels
  12. Mills, T.; Moody, K.; Rodden, K.: Providing world wide access to historical sources (1997) 0.05
    0.05106178 = product of:
      0.20424712 = sum of:
        0.20424712 = weight(_text_:java in 3697) [ClassicSimilarity], result of:
          0.20424712 = score(doc=3697,freq=2.0), product of:
            0.43718386 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06203383 = queryNorm
            0.46718815 = fieldWeight in 3697, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=3697)
      0.25 = coord(1/4)
    
    Abstract
    A unique collection of historical material covering the lives and events of an English village between 1400 and 1750 has been made available via a WWW enabled information retrieval system. Since the expected readership of the documents ranges from school children to experienced researchers, providing this information in an easily accessible form has offered many challenges requiring tools to aid searching and browsing. The file structure of the document collection was replaced by an database, enabling query results to be presented on the fly. A Java interface displays each user's context in a form that allows for easy and intuitive relevance feedback
  13. Maarek, Y.S.: WebCutter : a system for dynamic and tailorable site mapping (1997) 0.05
    0.05106178 = product of:
      0.20424712 = sum of:
        0.20424712 = weight(_text_:java in 3739) [ClassicSimilarity], result of:
          0.20424712 = score(doc=3739,freq=2.0), product of:
            0.43718386 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06203383 = queryNorm
            0.46718815 = fieldWeight in 3739, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=3739)
      0.25 = coord(1/4)
    
    Abstract
    Presents an approach that integrates searching and browsing in a manner that improves both paradigms. When browsing is the primary task, it enables semantic content-based tailoring of Web maps in both the generation as well as the visualization phases. When search is the primary task, it enables contextualization of the results by augmenting them with the documents' neighbourhoods. This approach is embodied in WebCutter, a client-server system fully integrated with Web software. WebCutter consists of a map generator running off a standard Web server and a map visualization client implemented as a Java applet runalble from any standard Web browser and requiring no installation or external plug-in application. WebCutter is in beta stage and is in the process of being integrated into the Lotus Domino application product line
  14. Pan, B.; Gay, G.; Saylor, J.; Hembrooke, H.: One digital library, two undergraduate casses, and four learning modules : uses of a digital library in cassrooms (2006) 0.05
    0.05106178 = product of:
      0.20424712 = sum of:
        0.20424712 = weight(_text_:java in 907) [ClassicSimilarity], result of:
          0.20424712 = score(doc=907,freq=2.0), product of:
            0.43718386 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06203383 = queryNorm
            0.46718815 = fieldWeight in 907, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=907)
      0.25 = coord(1/4)
    
    Abstract
    The KMODDL (kinematic models for design digital library) is a digital library based on a historical collection of kinematic models made of steel and bronze. The digital library contains four types of learning modules including textual materials, QuickTime virtual reality movies, Java simulations, and stereolithographic files of the physical models. The authors report an evaluation study on the uses of the KMODDL in two undergraduate classes. This research reveals that the users in different classes encountered different usability problems, and reported quantitatively different subjective experiences. Further, the results indicate that depending on the subject area, the two user groups preferred different types of learning modules, resulting in different uses of the available materials and different learning outcomes. These findings are discussed in terms of their implications for future digital library design.
  15. Mongin, L.; Fu, Y.Y.; Mostafa, J.: Open Archives data Service prototype and automated subject indexing using D-Lib archive content as a testbed (2003) 0.05
    0.05106178 = product of:
      0.20424712 = sum of:
        0.20424712 = weight(_text_:java in 2167) [ClassicSimilarity], result of:
          0.20424712 = score(doc=2167,freq=2.0), product of:
            0.43718386 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06203383 = queryNorm
            0.46718815 = fieldWeight in 2167, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=2167)
      0.25 = coord(1/4)
    
    Abstract
    The Indiana University School of Library and Information Science opened a new research laboratory in January 2003; The Indiana University School of Library and Information Science Information Processing Laboratory [IU IP Lab]. The purpose of the new laboratory is to facilitate collaboration between scientists in the department in the areas of information retrieval (IR) and information visualization (IV) research. The lab has several areas of focus. These include grid and cluster computing, and a standard Java-based software platform to support plug and play research datasets, a selection of standard IR modules and standard IV algorithms. Future development includes software to enable researchers to contribute datasets, IR algorithms, and visualization algorithms into the standard environment. We decided early on to use OAI-PMH as a resource discovery tool because it is consistent with our mission.
  16. Song, R.; Luo, Z.; Nie, J.-Y.; Yu, Y.; Hon, H.-W.: Identification of ambiguous queries in web search (2009) 0.05
    0.05106178 = product of:
      0.20424712 = sum of:
        0.20424712 = weight(_text_:java in 3441) [ClassicSimilarity], result of:
          0.20424712 = score(doc=3441,freq=2.0), product of:
            0.43718386 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06203383 = queryNorm
            0.46718815 = fieldWeight in 3441, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=3441)
      0.25 = coord(1/4)
    
    Abstract
    It is widely believed that many queries submitted to search engines are inherently ambiguous (e.g., java and apple). However, few studies have tried to classify queries based on ambiguity and to answer "what the proportion of ambiguous queries is". This paper deals with these issues. First, we clarify the definition of ambiguous queries by constructing the taxonomy of queries from being ambiguous to specific. Second, we ask human annotators to manually classify queries. From manually labeled results, we observe that query ambiguity is to some extent predictable. Third, we propose a supervised learning approach to automatically identify ambiguous queries. Experimental results show that we can correctly identify 87% of labeled queries with the approach. Finally, by using our approach, we estimate that about 16% of queries in a real search log are ambiguous.
  17. Croft, W.B.; Metzler, D.; Strohman, T.: Search engines : information retrieval in practice (2010) 0.05
    0.05106178 = product of:
      0.20424712 = sum of:
        0.20424712 = weight(_text_:java in 3605) [ClassicSimilarity], result of:
          0.20424712 = score(doc=3605,freq=2.0), product of:
            0.43718386 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06203383 = queryNorm
            0.46718815 = fieldWeight in 3605, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=3605)
      0.25 = coord(1/4)
    
    Abstract
    For introductory information retrieval courses at the undergraduate and graduate level in computer science, information science and computer engineering departments. Written by a leader in the field of information retrieval, Search Engines: Information Retrieval in Practice, is designed to give undergraduate students the understanding and tools they need to evaluate, compare and modify search engines. Coverage of the underlying IR and mathematical models reinforce key concepts. The book's numerous programming exercises make extensive use of Galago, a Java-based open source search engine. SUPPLEMENTS / Extensive lecture slides (in PDF and PPT format) / Solutions to selected end of chapter problems (Instructors only) / Test collections for exercises / Galago search engine
  18. Tang, X.-B.; Wei Wei, G,-C.L.; Zhu, J.: ¬An inference model of medical insurance fraud detection : based on ontology and SWRL (2017) 0.05
    0.05106178 = product of:
      0.20424712 = sum of:
        0.20424712 = weight(_text_:java in 4615) [ClassicSimilarity], result of:
          0.20424712 = score(doc=4615,freq=2.0), product of:
            0.43718386 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06203383 = queryNorm
            0.46718815 = fieldWeight in 4615, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.046875 = fieldNorm(doc=4615)
      0.25 = coord(1/4)
    
    Abstract
    Medical insurance fraud is common in many countries' medical insurance systems and represents a serious threat to the insurance funds and the benefits of patients. In this paper, we present an inference model of medical insurance fraud detection, based on a medical detection domain ontology that incorporates the knowledge base provided by the Medical Terminology, NKIMed, and Chinese Library Classification systems. Through analyzing the behaviors of irregular and fraudulent medical services, we defined the scope of the medical domain ontology relevant to the task and built the ontology about medical sciences and medical service behaviors. The ontology then utilizes Semantic Web Rule Language (SWRL) and Java Expert System Shell (JESS) to detect medical irregularities and mine implicit knowledge. The system can be used to improve the management of medical insurance risks.
  19. Kerr, M.: Using the Internet for business information : practical tips and hints (2004) 0.05
    0.04684681 = product of:
      0.18738724 = sum of:
        0.18738724 = weight(_text_:hints in 5505) [ClassicSimilarity], result of:
          0.18738724 = score(doc=5505,freq=2.0), product of:
            0.5128635 = queryWeight, product of:
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.06203383 = queryNorm
            0.36537448 = fieldWeight in 5505, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.03125 = fieldNorm(doc=5505)
      0.25 = coord(1/4)
    
  20. Cohen, D.J.: From Babel to knowledge : data mining large digital collections (2006) 0.05
    0.04684681 = product of:
      0.18738724 = sum of:
        0.18738724 = weight(_text_:hints in 2178) [ClassicSimilarity], result of:
          0.18738724 = score(doc=2178,freq=2.0), product of:
            0.5128635 = queryWeight, product of:
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.06203383 = queryNorm
            0.36537448 = fieldWeight in 2178, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              8.267481 = idf(docFreq=30, maxDocs=44421)
              0.03125 = fieldNorm(doc=2178)
      0.25 = coord(1/4)
    
    Abstract
    In Jorge Luis Borges's curious short story The Library of Babel, the narrator describes an endless collection of books stored from floor to ceiling in a labyrinth of countless hexagonal rooms. The pages of the library's books seem to contain random sequences of letters and spaces; occasionally a few intelligible words emerge in the sea of paper and ink. Nevertheless, readers diligently, and exasperatingly, scan the shelves for coherent passages. The narrator himself has wandered numerous rooms in search of enlightenment, but with resignation he simply awaits his death and burial - which Borges explains (with signature dark humor) consists of being tossed unceremoniously over the library's banister. Borges's nightmare, of course, is a cursed vision of the research methods of disciplines such as literature, history, and philosophy, where the careful reading of books, one after the other, is supposed to lead inexorably to knowledge and understanding. Computer scientists would approach Borges's library far differently. Employing the information theory that forms the basis for search engines and other computerized techniques for assessing in one fell swoop large masses of documents, they would quickly realize the collection's incoherence though sampling and statistical methods - and wisely start looking for the library's exit. These computational methods, which allow us to find patterns, determine relationships, categorize documents, and extract information from massive corpuses, will form the basis for new tools for research in the humanities and other disciplines in the coming decade. For the past three years I have been experimenting with how to provide such end-user tools - that is, tools that harness the power of vast electronic collections while hiding much of their complicated technical plumbing. In particular, I have made extensive use of the application programming interfaces (APIs) the leading search engines provide for programmers to query their databases directly (from server to server without using their web interfaces). In addition, I have explored how one might extract information from large digital collections, from the well-curated lexicographic database WordNet to the democratic (and poorly curated) online reference work Wikipedia. While processing these digital corpuses is currently an imperfect science, even now useful tools can be created by combining various collections and methods for searching and analyzing them. And more importantly, these nascent services suggest a future in which information can be gleaned from, and sense can be made out of, even imperfect digital libraries of enormous scale. A brief examination of two approaches to data mining large digital collections hints at this future, while also providing some lessons about how to get there.

Authors

Languages

Types

  • a 9373
  • m 2241
  • el 1010
  • x 591
  • s 554
  • i 168
  • r 116
  • ? 66
  • n 55
  • b 47
  • l 23
  • p 22
  • h 17
  • d 15
  • u 14
  • fi 10
  • v 2
  • z 2
  • au 1
  • ms 1
  • More… Less…

Themes

Subjects

Classifications