57
3.0 Introduction
Technology is transforming the nature of libraries and librarians and it continues
to exert a huge impact on the strategic directions of libraries in society. Nowadays, the
library services are changing from local collections to global resources provided on
demand via the most advanced networking technologies. Libraries in general and health
science libraries in particular are challenged to provide greater information access and
improved level of service, while coping with the pace of technological change and ever-
increasing budget pressure. As a result information professional librarians must be
computer literate and knowledgeable about information technology in order to render
quality services to patrons.
An attempt is made in this chapter to delineate the information technology skills
considered in this study. Information technology skills for the purpose of the present
study are as follows:
� Basic principles concerning automated systems, data standards, and
system analysis techniques including design and evaluation;
� Acquisition, use, and evaluation of information technologies;
� Integration of systems and technologies;
� Technological solutions for permanent access to electronic information;
� Applications of emerging areas of biomedicine computational biology and
health information, including electronic health care systems and records;
� Communication and information infrastructure including the Internet and
Web;
� Assess, select and apply current and emerging information tools and create
information access and delivery solutions;
� Apply expertise in databases and information analysis and synthesis to
improve information retrieval and use in the organization
58
3.1 Basic Principles Concerning Automated Systems, Data Standards, and System
Analysis Techniques
This section deals with basic principles of automated systems including, power of
source, feedback controls, and machine programming; library automation, integrated
library management systems including Easylib, LibSys, Nosa and Pars Azarakhsh
software packages; and information retrieval systems such as CDSISIS.
3.1.1 Basic Principles of Automated Systems
There are three basic building blocks of automation: (1) a source of power to
perform some action, (2) feedback controls, and (3) machine programming. Almost
without exception, an automated system will exhibit all these elements (Britannica CD,
1995).
3.1.1.1 Power Source
An automated system is designed to accomplish some useful action, and that
action requires power. There are many sources of power available, but the most
commonly used power in today's automated systems is electricity. The actions performed
by automated systems are generally of two types: (1) processing and (2) transfer and
positioning. In the first case, energy is applied to accomplish some processing operation
on some entity. In automated communications and information systems, the terms
transfer and positioning refer to the movement of data (or electrical signals) among
various processing units and the delivery of information to output terminals (printers,
video display units, etc.) for interpretation and use by humans. (Britannica CD, 1995)
3.1.1.2 Feedback Controls
Feedback controls are widely used in modern automated systems. A feedback
control system consists of five basic components: (1) input, (2) process being controlled,
(3) output, (4) sensing elements, and (5) controller and actuating devices.
The input to the system is the reference value, or set point, for the system output.
This represents the desired operating value of the output. The sensing elements are the
59
measuring devices used in the feedback loop to monitor the value of the output variable.
The purpose of the controller and actuating devices in the feedback system is to compare
the measured output value with the reference input value and to reduce the difference
between them. (Britannica CD, 1995)
3.1.1.3 Machine Programming
The programmed instructions determine the set of actions that is to be
accomplished automatically by the system. The program specifies what the automated
system should do and how its various components must function in order to accomplish
the desired result. (Britannica CD, 1995)
3.1.2 Library Automation
Library automation refers to the use of automatic machine or processing devices
in the libraries. The automation may be applied to library administrative activities, office
procedures, and delivery of library services to the users. (Biology online dictionary,
2006). According to Marimuthu and Paraman (2011, p. 11) Library automation means
not only entering and reading data on the computer, but the different functional areas of a
library also should be automated.
3.1.2.1 Integrated Library Management Systems
An integrated library Management system, or ILMS, is an enterprise resource
planning system for a library used to track items owned, orders made, bills paid, and
patrons who have borrowed. An ILMS is usually comprised of a relational database,
software to act on that database, and two graphical user interfaces (one for patrons, one
for staff). Most ILMS separate software functions into discrete programs called modules,
which are then integrated into a unified interface. Examples of modules include:
acquisitions (ordering, receiving, and invoicing materials), cataloging (classifying and
indexing materials), circulation (lending materials to patrons and receiving them back),
serials (tracking magazine and newspaper holdings), and the OPAC (New World
Encyclopedia, 2009). Examples of ILMSs are EasyLib, and Libsys in India and Pars
Azarakhsh and Nosa in Iran.
60
3.1.2.1.1 EasyLib Software
Easylib Library Software is one of the best Library Software available in India.
With nearly 400 Clients throughout India, it is now well proven software for Library
Automation. (Linkedin, 2011)
3.1.2.1.2 LibSys Software
LibSys is an integrated multiuser library management software, that caters to the
needs of an advanced library and information professionals. It provides a tree structure
system with each system comprising of several sub-systems having unmatchable depth in
functionality. It has a powerful and user-friendly Web-OPAC along with Windows-based
OPAC. It runs on various platforms such as WINDOWS (95/98/NT/2000/XP), UNIX
(various flavors), LINUXM, etc. Further, it adheres to standards such as MARC and
Z39.50 that makes it suitable for cooperative networking and resource sharing (Indian
Institute of Management-Calcutta, 2011).
3.1.2.1.3 Paras Azarakhsh and Nosa Software
Nosa and Pars Azarakhsh are commercial software available in Iran that are
programmed for data processing in both Persian and English languages and easy to look
after in terms of vendor support. (Davarpanah, 2003)
3.1.3 Information Retrieval Systems
An information retrieval system is a system that is capable of storage, retrieval,
and maintenance of information. Information in this context can be composed of text
(including numeric and date data), image, audio, video, and other multi-media objects.
(Kowalski, p. 2, 1997).
3.1.3.1 CDS/ISIS
CDS/ISIS (Computerized Documentation Systems/ Integrated Set of Information
Systems) is an Integrated Storage and Information retrieval System of United Nations
Educational Scientific and Cultural Organization (UNESCO), which is widely used
61
throughout the world especially in Latin America and Africa for creating metadata
management systems. There are hundreds of CDS/ISIS based application systems
managing bibliographical reference ensuring high quality content through the use of
built-in authority files, data entry guidelines and validations. (Kaloyanova, Betti, and
Keizer, 2008)
3.1.4 Bibliographic Utilities
An organization that maintains a database of bibliographic descriptions for use by
organizations and individuals for purposes of access, cataloging, circulation, and
management. These utilities grew out of cooperative cataloging projects in libraries,
allowing one library to catalog a book and other libraries to use that record in their
catalog. The utilities were able to exploit the information in the database to provide
interlibrary loan services to member libraries and to provide subscribers the ability to
search for works across many libraries' holdings. Two prominent utilities are OCLC and
RLIN (Pearce-Moses, 2005).
3.1.4.1OCLC (Online Computer Library Center)
OCLC is a nonprofit, membership, computer library service and research
organization dedicated to the public purposes of furthering access to the world’s
information and reducing information costs. More than 72,000 libraries in 170 countries
and territories around the world have used OCLC services to locate, acquire, catalog, lend
and preserve library materials. (OCLC, 2011) OCLC has products and services
3.1.4.2 LocatorPlus
LocatorPlus is the National Library of Medicine (NLM) online catalog available
at http://locatorplus.gov. LocatorPlus is continuously updated and includes:
� Computer files, and other materials in the Library's collections
� Holdings information for journals and other materials
� Links from catalog records to Internet resources, including online journals
62
� Circulation status information for materials, including those on-order or in-
process at the Library
LocatorPlus is one component of NLM's integrated library system, Voyager.
(National Library of Medicine, 2009)
3.1.5 Data Standards
Standards are mutually agreed-upon statements that help control an action or
product. Data standards promote the consistent recording of information and are
fundamental to the efficient exchange of information. They provide the rules for
structuring information, so that the data entered into a system can be reliably read, sorted,
indexed, retrieved, communicated between systems, and shared. They help protect the
long-term value of data. There are at least four data standards: data structure, data
content, data value, and data communication standards. (Visual Resources Association,
1997).
3.1.5.1 Data Structure Standards
Data structure standards are concerned with the definition of a record and the
relationship of the fields within it. The MARC (MAchine Readable Cataloging) format is
a good example, although it is really a hybrid of a data structure and an information
exchange standard. (Visual Resources Association, 1997).
3.1.5.2 Data Content Standards
Data content standards are the rules for how data are entered, for example
cataloging rules and syntax conventions. Examples of this are the Anglo-American
Cataloguing Rules (AACR2), Graphic Materials: Rules for Describing Original Items and
Historical Collections by Elisabeth Betz Parker, and the Art and Architecture Thesaurus
(AAT) Application Protocol, which suggests protocols for combining individual AAT
terms to form headings or strings for indexing. (Visual Resources Association, 1997).
63
3.1.5.3 Data Value Standards
Date value standards usually take the form of controlled vocabularies, including
subject specific-terminologies and authorities for names and places (Visual Resources,
1997). Well-known examples include the Library of Congress Subject Headings (LCSH),
and Medical Subject Headings (MeSH). (Visual Resources Association, 1997)
3.1.6 Library of Congress Subject Headings (LCSH)
The LCSH schema was originally designed as a controlled vocabulary for
describing books and other materials in the Library of Congress (LOC) collection. LCSH
are a list of pre-coordinated terms that are proposed by catalogers to represent all items in
the Library of Congress (LOC) and are based on literary warrant. Besides representing
materials in library collections, the LCSH have a dual purpose of providing information
seekers access to information on any subject and to provide material on related topics
(Wynar, 1985).
3.1.7 Medical Subject Headings (MeSH)
MeSH is the National Library of Medicine's controlled vocabulary thesaurus. It
consists of sets of terms naming descriptors in a hierarchical structure that permits
searching at various levels of specificity. MeSH descriptors are arranged in both an
alphabetic and a hierarchical structure. The MeSH thesaurus is used by NLM for
indexing articles from 5,400 of the world's leading biomedical journals for the
MEDLINE/PubMED database. It is also used for the NLM-produced database that
includes cataloging of books, documents, and audiovisuals acquired by the Library.
(National Library of Medicine, 2011)
3.1.8 Z39.50
Z39.50 is a standard that defines a protocol to support the searching and retrieval
of information across networks. This information can be in a whole range of formats such
as full text documents, bibliographic data, images or multimedia. For this reason, Z39.50
is often thought of as a library standard. Although not the only sector using Z39.50 (it is
64
also used by museums, government and those working with geospatial data records), the
library community is by far the oldest and most established user. (Taylor, 2003)
3.1.9 Resource Description Framework (RDF)
RDF is a general framework for describing metadata of Internet resources and for
processing this metadata; it is a standard of World Wide Web Consortium (W3C). RDF
supports the interoperability between applications that exchange machine-understandable
information on the web. (Xing, Dikaiakos, Yang, Sphyris, and Eftichidis, 2005)
3.1.10 Dublin Core (DC)
Dublin Core Metadata Element Set (DCMES) is a standard metadata schema for
describing electronic records, documents and web resources. Without metadata, searching
the Internet or our corporate repositories would be similar to looking for a needle in a
haystack. Dublin Core (DC) metadata was developed by the Dublin Core Metadata
Initiative (DCMI). DCMI had its beginnings in 1995 in Dublin, Ohio, where the first of
the DC series workshops was held. Workshop participants came together to address the
pressing need for development in the description, organization, and discovery of
networked information resources (DCMI, 2001). This led to preparing 15 metadata
elements called Dublin Core Metadata Element Set (DCMES) to enrich the descriptive,
semantic and interoperability aspects of web resources. These elements are: title, creator,
subject, description, publisher, contributor, date, type, format, identifier, source,
language, relation, coverage, and rights, respectively (Saadat Alijani and Jowkar, 2009)
3.1.11 MARC (Machine Readable Cataloguing)
MARC is often described as an international standard that is used for any type of
information material bibliographic description. The purpose of that format is the
distribution of catalogue data using new computer technology and networks. While
MARC has been employed in developed countries for decades, it is only comparatively
recently that computer uptake and the extension of the internet in developing countries
has meant that MARC is being used (Ghaebi, Shamsbod and Karimi-Mansoorabad,
2010)
65
3.1.12 AACR2 (Anglo- American Cataloguing Rules)
AACR2 is an internationally accepted standard for descriptive cataloging. It
contains rules for describing and providing access to all types of library materials
including books, serials, computer files, maps, music, motion pictures, etc., through
library catalogs. AACR2 is also a standard for structuring catalogs with headings and
references to provide links between items with similar or related characteristics. (El-
Sherbini and Klim, 2004)
3.2 System Analysis
Systems analysis is an explicit formal inquiry carried out to help someone
(referred to as the decision maker) identify a better course of action and make a better
decision than he might otherwise have made. The characteristic attributes of a problem
situation where systems analysis is called upon are complexity of the issue and
uncertainty of the outcome of any course of action that might reasonably be taken.
Systems analysis usually has some combination of the following: identification and re-
identification) of objectives, constraints, and alternative courses of action; examination of
the probable consequences of the alternatives in terms of costs, benefits,
and risks; presentation of the results in a comparative framework so that the decision
maker can make an informed choice from among the alternatives. (Web Dictionary of
Cybernetics and Systems, 2000)
3.3 Skills in Acquisition, Use, and Evaluation of Information Technologies
One of the aims of evaluating information technologies is to address the added
value and increased quality provided by the provision of them. Here transaction log
analysis and electronic resource management/assessment systems are discussed as means
of evaluating electronic resource usage in libraries.
3.3.1 Transaction Log Analysis
A transaction log is an electronic record of interactions that have occurred between
a system and users of that system. These log files can come from a variety of computers
66
and systems (Websites, OPAC, user computers, blogs, listserv, online newspapers, etc.),
basically any application that can record the user – system – information interactions
(Jansen, Taksa, and Spink, 2009). Transaction log analysis is the methodological
approach to studying online systems and users of these systems. Peters (1993) defines
transaction log analysis as the study of electronically recorded interactions between
persons who search for information found in those systems.
3.3.2 Electronic Resource Assessment Systems (ERAS)
Usage analysis, in contrast to citation analysis, is an emerging area of bibliometric
research. Although impact factor (IF) is a widely accepted metric for determining the
value of a journal, ISI coverage of journals is not comprehensive and is only a segment of
the journals published in each discipline. Electronic resource assessment systems are used
for usage analysis. They can provide libraries with a different and potentially more
complete perspective on the scholarly impact of electronic resources than citation data
alone can provide (Xu, 2010). Since 1998, the library community has been requesting
that electronic resource vendors comply with the International Coalition of Library
Consortia (ICOLC) guidelines for statistical measures of usage.
3.3.2.1 Counting Online Usage of Networked Electronic Resources (COUNTER)
In 2002, the project Counting Online Usage of Networked Electronic Resources
(COUNTER) Codes of Practice emerged as an organised effort to standardise how the
elements are counted in database vendors usage report. Building upon the previous
initiatives, particularly the ICOLC guidelines, COUNTER aims to develop a Code of
Practice that will ensure that the tools for measuring electronic resources will provide
results that are consistent, credible, and compatible (COUNTER, 2009). With the project
COUNTER Codes of Practice, vendors can provide libraries with reliable, accurate, and
standardised usage reports. Several private sector vendors developed commercial
electronic resource assessment systems, such as USAS (Ulrich’s Serials Analysis
System), ScholarlyStats, Journal Use Reports (JUR), 360 Counter, and Scholarly iQ,
which complemented the initiative of homegrown systems (Xu, 2010).
67
3.3.2.2 Ulrich’s Serials Analysis System (USAS)
USAS was developed by R R Bowker in January 2003. USAS is the first
evaluative service to provide collection analysis criteria specifically for serials. USAS
allows libraries to upload their current print and electronic serials and then analyse them
according to Ulrich’s subject categories and evaluative data for making collection
decisions. (Serials Solutions, 2010).
3.3.2.3 ScholarlyStats
In January 2006, MPS Technologies, a division of Macmillan Publishers Ltd,
introduced ScholarlyStats to the library community (Collins, 2006). ScholarlyStats is a
web-based portal, which provides libraries with online access to their aggregated
COUNTER (Counting Online Usage of Networked Electronic Resources) reports for
journals and databases by streamlining the data collection and consolidation process (Xu,
2010).
3.3.2.4 Journal Use Reports (JUR)
In 2006, Thomson Scientific launched JUR, which brings COUNTER-compliant
use statistics, publication activity, and citation metrics into a single interface for an
institution (Brown and Dumouchel, 2007; Chernyi, 2009). JUR is comprised of several
data sets, including journal influence from JCR (Journal Citation Report), institutional
publication and citation activity from Web of Science that can tell in which journals the
authors at a local institution publish and which journals they rely on most to support their
published research, and COUNTER-compliant usage data (Liu, 2007).
3.3.2.5 360 Counter
In November 2007, Serials Solutions, a business unit of ProQuest Information and
Learning, announced the launch of the 360 Counter e-resource assessment service. 360
Counter helps libraries evaluate the usage and cost of e-resources to optimise the value of
electronic collections (Serials Solutions, 2007). Like ScholarlyStats, 360 Counter is a
68
web-based portal that offers libraries online access to their aggregated COUNTER
reports for journals and databases (Xu, 2010).
3.3.2.6 Scholarly iQ
In June 2009, the Association for Computing Machinery (ACM) announced an
enhanced web analytics platform for library communities to create custom reports that
comply with COUNTER standards. The new platform, provided by Scholarly iQ
(www.scholarlyiq.com/), enables librarians to access usage statistics and receive timely
accurate COUNTER usage reports for measuring online usage of scholarly journals (Xu,
2010).
3.4 Integration of Systems and Technologies
Integration of systems and technologies refers to the melding of systems and new
technologies to form more capable systems that are intended to take on additional tasks,
exhibit improved performance or enhance existing systems (California State University,
2011). A typical academic library will likely offer at least half a dozen different Web-
based applications, each providing a specialized service. Each of these applications may
come from a different vendor. Web-based services offered by a library might include the
online catalog of its automation system, an interlibrary loan or document delivery system,
an OpenURL-based reference linking environment, a meta search interface, electronic
reserves, a virtual reference service, remote patron authentication, and/or local digital
library collections, as well as a plethora of subscriptions to citation databases, full-text
resources, and electronic journals. Having such a variety of services and resources
increases the complexity of the overall environment, and makes it harder to integrate
everything into an environment that appears simple to the users (Breeding 2004).
3.4.1Content Development
Content Development or Website Content Development is the act of preparing
and developing your website's content to ensure that it is properly setup, crawled, and
indexed by the search engines. Content Development is a major element of SEO (Search
69
Engine Optimization) and requires close attention to ensure the success of your website
(Hudson Horizons: Web Site design and development company, 2011).
3.4.1.1 Library Portals
Library Portals are a subset of Web portals and serve specific research
communities. Available for some time now, library portals typically provide a gateway to
an institution’s resources by listing them for users and creating a direct link to the native
interface of each resource. Such listings are available on most library Web sites today,
although many sites provide only alphabetic listings. (Sadeh and Walker, 2003)
3.4.1.2 Intranet
A private network based on the TCP/IP protocol suite and designed to provide
access to information resources within a company, university, or other organization.
Designed to look much like a site on a private World Wide Web (WWW) and based on
the same protocols, an intranet supports familiar client/server software such as browsers
and e-mail (Your dictionary, 2011). Several researchers reported application of Intranet
in knowledge and content management, resource sharing, and reference services
(Delaney, 1999; Engel and Bierman, 2006; Mphidi and Snyman, 2004; Pandian,
Jambhekar, and Karisiddappa, 2002; Robbins, Engel and Bierman, 2006).
3.4.1.3 VLE (Virtual Learning Environment)
VLEs began to appear in the second half of the 1990s. They were initially non-
web-based – and some still are. Gradually, the Web is taking over as the dominant form,
for obvious reasons. The main advantage in developing courses on the Web, however, is
the ability to link resources in from elsewhere on an intranet or the Internet. One of the
leading VLEs in Higher Education (HE) at the present time is WebCT (www.webct.com)
which was developed originally in the University of British Columbia and gained a
significant foothold in the academic market initially in North America because of its
understanding of the needs of university courses (Maccoll, 2001). In general, these
applications offer a common range of functions, which comprise the following basic set:
70
� e-mail to allow broadcasting of messages to the course group, and
individual
� dialogue between students and tutors;
� alerting;
� calendar/scheduling, to accommodate timetabling information;
� a structure to facilitate course and module description;
� a facility to permit other web resources to be linked in;
� assessment tools (typically shells which permit multiple choice tests to
be developed).
3.4.1.4 Learning Objects (LO)
Learning objects can be defined as: “Any digital resource that can be reused to
support learning”. Because learning objects are in digital formats, they are deliverable
over the Internet and therefore accessible by any number of people. A learning object can
be a single file such as an animation, a video clip, a discrete piece of text or URL, or it
could be a collection of contextualised files that make up a learning sequence. It is a
digital resource that can be identified, tracked, referenced, used and reused for a variety
of learning purposes (Wiely, 2002).
3.4.1.5 Content Management Systems (CMSs)
A content management system (CMS) offers a way to manage large amounts of
web-based information that escapes the burden of coding all of the information into each
page in HTML by hand. Content management systems are important to libraries because
the sheer mass of their public web presence has reached the point where maintenance is a
problem. Many libraries also want to customize their look and feel to emphasize
consistency and branding. That can be very difficult when it requires multiple authors to
agree on fonts, formats, logos, and other limitations on their creativity and individuality.
A CMS assists conformity by providing a centrally managed system for displaying the
71
content, which still remains under the control of library staff with appropriate subject
expertise (Seadle, 2006).
3.4.1.6 Referencing Software
Referencing Software (sometimes referred to as Bibliographic Management
Software) is designed to help researchers store the references which they have located,
and then cite those references in the essay, paper, thesis or book which they are writing.
Examples are Endnote, RefWorks, and Zotero. (The University of Queensland, 2011)
3.4.1.7 Digital Libraries
A digital library (DL) is a collection of services and “information objects” that are
available digitally. “Information objects” can be defined as anything in a digital format
such as books, journal articles and sounds, since DLs organise and present information
objects to users, and support them in dealing with these objects (Leiner, 2005).
3.4.1.8 Optical Character Recognition (OCR)
Optical Character Recognition, or OCR, is a technology that is used to convert
different types of documents, such as scanned paper documents, PDF files or images
captured by a digital camera into editable and searchable data (ABBYY, 2011).
3.4.1.9 Digital Asset Management System (DAMS)
Digital Asset Management System (DAMS) is an integrated suite of
infrastructure components used to capture, catalog, store, and manage digital assets, and
to expose those assets to creative tools for producing video, audio, Web, and print
content. While, DAMS is related to the market segments of document management, Web
content management, and integrated library systems, it is viewed as an infrastructure
component that supports these other environments. A DAMS infrastructure can ingest
digital assets, store and index assets for easy searching, retrieve assets for use in many
environments, and manage the rights associated with those assets (McCord, 2005).
72
3.4.1.10 Metadata Systems
Metadata are data about data. The term came into being with the appearance of
electronic resources and it initially referred to standards that assisted in identifying,
describing and locating electronic resources. This concept continued to evolve as it was
applied to a variety of “non-traditional” formats, resulting in proliferation of metadata
schemes. Metadata standards existing today range from very complex to very simple.
Relative simplicity or complexity of metadata standards depends in large part on the
resources for which they were created and the depth of description that is deemed
necessary to make these resources accessible. (El-Sherbini and Klim, 2004)
3.4.1.11 Meta Search Systems
Meta searching, also known as integrated searching, simultaneous searching,
cross-database searching, parallel searching, and broadcast searching, refers to a process
in which a user submit a query simultaneously to numerous information resources. The
resources can be heterogeneous in many respects; their location, the format of the
information that they offer, the technologies on which they draw, the types of materials
that they contain, and more. The user’s query is broadcast to each resource, and results
are returned to the users (Sadeh, 2005). Metalib and Webfeat are examples of meta search
systems.
3.4.1.11.1 MetaLib
MetaLib is a federated search system developed by Ex Libris Ltd. MetaLib
conducts simultaneous searches in multiple, and often heterogeneous, information
resources such as library catalogs, journal articles, newspapers and selected quality
Internet resources. (Ex Libris Metalib, 2011)
3.4.1.11.2 WebFeat
WebFeat launched the original federated search engine in 1998 to let libraries
search any or all of their databases at the same time. Over 16,000 libraries currently use
WebFeat - from small schools to the largest public and academic libraries in the world.
73
WebFeat's patented federated search technology makes it compatible with virtually any
databases, WebFeat supports over 9,000 databases. (WebFeat, 2011)
3.4.1.12 Virtual Reference Services
Reference and User Services Association (RUSA) defines virtual reference
service as: “reference service initiated electronically, often in real-time, where patrons
employ computers or other internet technology to communicate with reference staff,
without being physically present. Communication channels used frequently in virtual
reference include chat, video conferencing, voice over IP, co-browsing, e-mail, and
instant messaging (RUSA, 2004).
3.4.1.13 Electronic Reserve Systems
Some educational institutions are now providing electronic reserve systems that
allow storage of electronic versions of materials that students may retrieve on a computer
screen, and from which they may print a copy for their personal study. When materials
are included as a matter of fair use, electronic reserve systems should constitute an ad hoc
or supplemental source of information for students, beyond a textbook or other materials.
(Copyright Crash Course, 2001).
3.4.1.14 Open-URL Link Resolver
Open-URL stands for Open Universal Resource Locator. Open-URL link resolver
technology is designed to remove obstacles from users searching electronic resources. It
enables the searcher to go directly from an individual citation to the full text referred to
by that citation in one mouse click. For example, if a student finds a citation in the
Modern Language Association Bibliography, he/she can click on the open-URL icon and
be connected to his/her destination, the full text article (Yi and Herlihy, 2007).
3.4.1.15 Remote Patron Authentication (RPA)
Today, libraries have been charged with the responsibility of controlling Internet-
based access to restricted resources. Patrons are coming to expect access to library
resources whether they are inside the library or at some remote location.
74
Remote Patron Authentication delivers real-time patron authentication against
specific library systems, requests redirection to restricted resources, and requests
authentication identification to online content and database providers. Referring URL,
SharedID/Password, Proxy Server, EZproxy are examples of remote access method.
3.4.1.16 Programming Languages
A programming language is a set of rules that provides a way of instructing the
computer to perform certain operations (Leon and Leon, 1999, p. 12.1).
Programming languages are said to be lower or higher, depending on whether they are
closer to the language the computer itself uses (lower which means 0s and 1s) or to the
language that people use (higher, which means more English like). Perl, Java, C and C++
are examples of programming languages.
3.4.1.16.1 Perl
Perl is a popular scripting language that runs on a wide variety of platforms,
including UNIX and Windows. Perl is open source, easily integrated into Web servers for
Common Gateway Interface (CGI) , easy to learn, and supports a large library of utilities.
(Your dictionary, 2011). Now Perl is used for a wide range of tasks including system
administration, web development, network programming, bioinformatics, and
development. (Your dictionary, 2011)
3.4.1.16.2 JAVA
Java is a programming language and computing platform first released by Sun
Microsystems in 1995. It is the underlying technology that powers state-of-the-art
programs including utilities, games, and business applications. Java runs on more than
850 million personal computers worldwide, and on billions of devices worldwide,
including mobile and TV devices. There are lots of applications and websites that won't
work unless you have Java installed, and more are created every day. Java is fast, secure,
and reliable. From laptops to datacenters, game consoles to scientific supercomputers,
cell phones to the Internet, Java is everywhere. (Java, 2011)
75
3.4.1.16.3 C++
An object-oriented version of C created by Bjarne Stroustrup at AT&T Bell
Laboratories in 1986. In C, "C++" means "add 1 to C." C++ is the basis of the Java
language. As one of the most popular programming languages ever created, C++ is
widely used in the software industry. Some of its application domains include systems
software, application software, device drivers, embedded software, high-performance
server and client applications, and entertainment software such as video games. (Tech
dictionary, 2011)
3.4.2 Federated Search Engines
Conventional search engines like Google provide access to Web information that
can be acquired easily by crawling hyperlinks. However, a large amount of information
cannot be copied arbitrarily by conventional search engines. This type of hidden
information, which is very valuable, can only be accessed via an alternative search model
other than the centralized retrieval model used by conventional search engines. Federated
search provides access to the hidden information by providing a single interface that
connects to multiple source-specific search engines (Si, 2007). Google Scholar, and
Scirus, are federated search engines among others.
3.4.2.1 Google Scholar
Google Scholar (http://scholar.google.com/) is a search engine that provides a
simple way to broadly search for scholarly literature. From one place, one can search
across many disciplines and sources: articles, theses, books, abstracts and court opinions,
from academic publishers, professional societies, online repositories, universities and
other web sites. Google Scholar helps to find relevant work across the world of scholarly
research.
Google Scholar aims to rank documents the way researchers do, weighing the full
text of each document, where it was published, who it was written by, as well as how
often and how recently it has been cited in other scholarly literature. (Google Scholar,
2011)
76
3.4.2.2 Scirus
Scirus (http://www.scirus.com/) is the most comprehensive science-specific
search engine on the Internet. Driven by the latest search engine technology, Scirus
searches over 410 million science-specific Web pages, enabling researchers to quickly:
� Pinpoint scientific, scholarly, technical and medical data on the Web.
� Find the latest reports, peer-reviewed articles, patents, pre prints and journals
that other search engines miss.
� Offer unique functionalities designed for scientists and researchers.
Scirus has proved so successful at locating science-specific results on the Web
that the Search Engine Watch Awards voted Scirus 'Best Specialty Search Engine' in
2001 and 2002 and 'Best Directory or Search Engine Website' WebAward from Web
Marketing Association in 2004, 2005, 2006 and 2007. (Scirus, 2011)
3.4.2.3 Meta Search Engines
Unlike search engines, Meta Crawlers do not crawl the web themselves to build
listings. Instead, they allow searches to be sent to several search engines all at once. The
results are then blended together onto one page. Below are some of the major Meta
Crawlers. (Sherman, 2005)
3.4.2.4 Dogpile
Dogpile (http://www.dogpile.com) is a popular Meta search site owned by
InfoSpace that sends a search to a customizable list of search engines, directories, and
specialty search sites. Then displays results from each search engine individually.
(Sherman, 2005)
3.4.2.5 Vivisimo
After entering a search term, Vivisimo (http://www.Vivisimo.com) will not only
pull back matching responses from major search engines but also automatically organize
the pages into categories. It is slick and easy to use. (Sherman, 2005)
77
3.4.2.6 Kartoo
Kartoo (http://www.kartoo.com) is a Meta search site that shows the results with
sites being interconnected by keywords. (Sherman, 2005)
3.4.2.7 Mamma
Founded in 1996, Mamma (http://www.mamma.com) is one of the oldest Meta
search engines on the web. Mamma searches against a variety of major crawlers,
directories and specialty search sites. (Sherman, 2005)
3.4.3 RFID (Radio Frequency IDentification)
RFID stands for Radio Frequency IDentification, and in essence consists of a
miniature transmitter or receiver incorporated into a device known as a tag in a movable
item, which can be located by a receiver or transmitter which may be stationary or also
movable depending on the application. RFID technology is primarily intended to reveal a
current change in the library user service from the semi-automated to the fully-automated
mode (Chan and Ying, 2005). It is a fast growing technology used in libraries for
enhanced circulation capabilities, better inventory control, reliability, minimising theft of
documents, and provides batch access and storage of mass data.
3.4.4 Web-Based Knowledge Management
Knowledge management (KM) is viewed as a “process about acquisition,
creation, packaging, application and reuse of knowledge” Knowledge management is an
increasingly important source of competitive advantage for organizations. Hundreds of
millions of users can now access several billion documents on the web, and even larger
data sets reside in organizations, intranets and web-accessible databases (the so-called
deep web). Today web based knowledge management is playing a very important role in
big libraries. It has become more vital along-with the development of knowledge
economy (Rah, Gul, and Wani, 2010).
78
3.4.4.1 Web 2.0 Tools
The term “Web 2.0” was coined by O’Reilly Media at a conference in 2004 and it
has become the mechanism to refer to the next generation web. Web 2.0 is a knowledge-
oriented environment where human interactions generate content that is published and
used through network applications in a service-oriented architecture. Wikis, RSS, SNS,
Blogs and Mashups are all seen as components of Web 2.0. This variety of
communication tools available on Web 2.0 enables the introduction of a new KM model,
also called KM 2.0 or Enterprise 2.0 (Kakizawa, 2007; McAfee, 2006; Shimazu and
Koike, 2007 cited in Razmerita, Kirchner, and Sudzina, 2009, p. 1022). This model
tries to better harness the use of collective intelligence, and thus accelerates the
distribution of information.
3.4.4.1.2 Wikis
Wiki is a collaborative website whose content can be edited by anyone who has
access to it. It is a web application that allows users to add content, as on an Internet
forum, but also refers to the collaborative software used to create such a website. Wikis
are a knowledge management tool. This new tool allows for content to be worked on
collaboratively and produced at the same time. Put another way, it’s a tool for knowledge
creation (collaboration) and a tool for sharing explicit knowledge (content management)
rolled up into one (Rah, Gul, and Wani, 2010).
3.4.4.1.3 Blogs
Blog as a "first of the social software" which is also described as "chronological
social web site". Blogs have become a very popular, easy and social platform to exchange
information, and gradually are becoming a more acceptable reference sources even in
peer reviewed publications. There are many blogs on the Web offering news and trends
in the specific fields of human knowledge (Zarea Gavgani and Visha Mohan, 2008).
79
3.4.4.1.4 Really Simple Syndication (RSS)
RSS gives the possibility of being informed about new items added to the
interested site. Librarians can utilize the RSS feeds to catch up trends and changes of
desired medical specialty, and simultaneously to give the same opportunity to user
community to keep themselves up-to-date. Basically, blogs are not created by group of
people as wikis are created. But it is a collaborative source because readers can comment
and publish their own ideas about the topic, or ask more information (Zarea Gavgani and
Visha Mohan, 2008).
3.4.4.1.5 SNS (Social Networking Services)
SNS user-generated content sites that provide for some means, by which users
can connect and communicate with one another. Sites that offer a means of connection
can be either business-related (e.g., LinkedIn) or social in nature (e.g., MySpace, Orkut,
Facebook). Other networks combine both purposes (e.g., Ryze). Members can join user-
generated content sites through either self-registration or an existing offline connection; a
valid e-mail address is required in both cases (Murchu, Brestlin, and Decker 2004 cited in
Gangadharbatla, 2010).
3.4.4.1.6 Mashups
Originally the term used in Pop music by artists and disk jockeys when two songs
were remixed and played at the same time, Web experts have borrowed the term when
two or more software tools are merged. The resulting new tool provides an enriched Web
experience for end users. Like other social software, mashups are freely accessible. Prior
to the social software movement, only programming experts with training in C++ or
Visual Basic could publish complex Web sites. In contrast, a simple mushup can be
created for free in less than 15 minutes; expert technical techniques are not required
(Cho, 2007).
80
3.4.4.2 Web 3.0
Berners-Lee et al. (2001 in Sure and Studer, 2005, p.190) describe Web 3.0 or
semantic web as: “an extension of the current web in which information is given well-
defined meaning, better enabling computers and people to work in cooperation”. The key
enabler of the Semantic Web is the need of many communities to put machine-
understandable data on the web which can be shared and processed by automated tools as
well as by people. Machines should not just be able to display data, but rather be able to
use it for automation, integration and reuse across various applications.
3.4.4.2.1 Ontology
In philosophy, ontology is the study of the kinds of things that exist.” In Artificial
Intelligence (AI), the term ontology has largely come to mean one of the two related
things. First of all, ontology is a representation vocabulary, often specialized to some
domain or subject matter. In its second sense, the term ontology is sometimes used to
refer to a body of knowledge describing some domain, typically a commonsense
knowledge domain, using a representation vocabulary. Without ontologies, or the
conceptualizations that underlie knowledge, there cannot be a vocabulary for representing
knowledge. Thus, the first step in devising an effective knowledge representation system,
and vocabulary, is to perform an effective ontological analysis of the field, or domain
(Chandrasekaran, Josephson, and Benjamins, 1999).
3.4.4.2.2 Taxonomy
The term taxonomy is used with different meanings in different contexts. In the
domain of knowledge organization, taxonomy refers to listing of topics or categories
often used to provide a structured navigational path in a content collection. National
Information Standards Organization (National Information Standards Organisation, 2005)
defined taxonomy as a collection of controlled vocabulary terms arranged in a
hierarchical order.
81
3.4.5 Intelligent Agents
In computer science, intelligent agents often refer to software systems that are
able to take actions towards their goals without human intervention. An intelligent agent
may have the ability of reasoning, planning, learning and collaborating with users or
other agents in various environments. The intelligent agent technology has been adopted
in many areas, including industrial control, medical diagnosis, stock trading, personal
assistance, games, scientific discovery and information retrieval, etc. (Liu, 2011)
3.5 Skills in Technological Solutions for Permanent Access to Electronic
Information
Permanent Access and Archiving are of paramount importance. As with
traditional print materials, it is critical to libraries and the constituents they serve that
permanent archival access to electronic information be available, especially where that
information exists only in electronic form (International Coalition of Library Consortia
[ICOLC], 2004).
3.5.1 Digital Archiving and Preservation
Digital preservation is the set of processes and activities that ensure continued
access to information and all kinds of records, scientific and cultural heritage existing in
digital formats. This includes the preservation of materials resulting from digital
reformatting, but particularly information that is born-digital and has no analog
counterpart. In the language of digital imaging and electronic resources, preservation is
no longer just the product of a program but an ongoing process. In this regard the way
digital information is stored is important in ensuring its longevity. The long-term storage
of digital information is assisted by the inclusion of preservation metadata (Leiner,,
2005).
82
3.6 Skills in Applications of Emerging Areas of Biomedicine, Computational Biology
and Health Information, Including Electronic Health Care Systems and Records
Free Dictionary (2011) defines biomedicine as the branch of medical science that
deals with the ability of humans to tolerate environmental stresses and variations, as in
space travel and as the application of the principles of the natural sciences, especially
biology and physiology, to clinical medicine. Computational biology or sometimes called
bioinformatics is the field of biology concerned with the development of techniques for
the collection and manipulation of biological data, and the use of such data to make
biological discoveries or predictions. (Medical Glossary, 2011).
3.6.1 Computational Biology (Bioinformatics) Databases
Obviously before analysing any of DNA and protein sequences, they must first be
located and retrieved. Sequences are stored in vast databases world-wide. These
databases are usually either nucleotide (DNA) databases or protein databases.
By far the biggest nucleotide databases are the GenBank in Bethesta USA, the
EMBL (European Molecular Biology Laboratory) in Heidelberg Germany and the DNA
Data Bank of Japan (DDBJ) in Mishima. Each of the three collects a portion of the total
reported sequence data and changes it with the other on a daily basis. Protein databases
available include the SWISS-PROT database, maintained collaboratively by the EMBL
Data Library and Amos Bairoch of the University of Geneva and the Protein
Identification Resource (PIR) in Washington, D.C. (Vaughan and Pembroke, 2011).
3.6.2 Electronic Health Records
Electronic health records are also known as electronic medical records, electronic
patient records, and computerized patient records. These are computerized systems for
keeping all health information in one place. The personal information is recorded in
digital format and can be shared between multiple health care providers and across health
care settings (Weinberger, 2011).
83
3.6.3 Medical Informatics Standards
The need for standards in medical informatics is motivated by the same reasons
that standards are needed in other areas. Standards are needed in collection, manipulation
and transmission of information in medical informatics systems (Berg, 2005).
3.6.3.1 Unified Medical Language System (UMLS)
The purpose of NLM's Unified Medical Language System (UMLS) is to facilitate
the development of computer systems that behave as if they "understand" the meaning of
the language of biomedicine and health. To that end, NLM produces and distributes the
UMLS Knowledge Sources (databases) and associated software tools (programs) for use
by system developers in building or enhancing electronic information systems that create,
process, retrieve, integrate, and/or aggregate biomedical and health data and information,
as well as in informatics research. By design, the UMLS Knowledge Sources are multi-
purpose. They are not optimized for particular applications, but can be applied in systems
that perform a range of functions involving one or more types of information, e.g., patient
records, scientific literature, guidelines, public health data (National Library of Medicine,
2006).
3.6.3.2 RxNorm
RxNorm, a standardized nomenclature for clinical drugs and drug delivery
devices, is produced by the National Library of Medicine (NLM). In this context, a
clinical drug is a pharmaceutical product given to (or taken by) a patient with a
therapeutic or diagnostic intent. In RxNorm, the name of a clinical drug combines its
ingredients, strengths, and/or form. Because every drug information system that is
commercially available today follows somewhat different naming conventions, a
standardized nomenclature is needed for the smooth exchange of information, not only
between organizations, but even within the same organization.
The goal of RxNorm is to allow various systems using different drug
nomenclatures to share data efficiently at the appropriate level of abstraction ( National
Library of Medicine, 2011).
84
3.6.3.3 SNOMED CT (Systematized Nomenclature of Medicine--Clinical Terms)
SNOMED (Systematized Nomenclature of Medicine--Clinical Terms) is a
comprehensive clinical terminology, originally created by the College of American
Pathologists (CAP) and, as of April 2007, owned, maintained, and distributed by
the International Health Terminology Standards Development Organisation (IHTSDO), a
not-for-profit association in Denmark. The CAP continues to support SNOMED CT
operations under contract to the IHTSDO and provides SNOMED-related products and
services as a licensee of the terminology (National Library of Medicine, 2006).
3.6.3.4 DICOM (Digital Imaging and Communication in Medicine)
DICOM is a global Information-Technology standard that is used in virtually all
hospitals worldwide. Its current structure, which was developed in 1993, is designed to
ensure the interoperability of systems used to Produce, Store, Display, Process, Send,
Retrieve, Query or Print medical images and derived structured documents as well as to
manage related images (DICOM, 2005).
3.6.3.5 H7 (Health Level 7)
HL7 stands for Health Level 7, and is the highest level of the ISO communication
model. It is a standard for data interchange. The model’s purpose is to archive OSI (Open
Systems Interconnections), a way to get different systems to work together. The OSI
model is not specific to medical informatics. HL7 is specific to the health care domain
(Berg, 2005).
3.6.4 Telemedicine and Hospital Information Systems
Telemedicine is the use of medical information exchanged from one site to
another via electronic communications to improve patients' health status. Closely
associated with telemedicine is the term "telehealth," which is often used to encompass a
broader definition of remote healthcare that does not always involve clinical services.
Videoconferencing, transmission of still images, e-health including patient portals,
remote monitoring of vital signs, continuing medical education and nursing call centers
85
are all considered part of telemedicine and telehealth (American Telemedicine
Association, 2011).
3.6.4.1 Mercury
Mercury is a hybrid telemedicine system offering a suite of tools for enabling
sustainable telemedicine program across geographies it provides data insertion and
retrieval, Electronic Medical Record (EMR) building and it allows upload/download of
wide range of file formats such as images, audio/video and documents. This makes it
possible to store/view all kinds of data in EMR (Web telemedicine distributed, 2005) .
3.6.4.2 Sanjeevani
'Sanjeevani' can be classified as a hybrid model of integrated telemedicine
solution that can be used in a store and forward as well as real time mode. Sanjeevani is a
desktop-based application and is designed to facilitate doctor-to-doctor consultation with
the provision of real time patient visualisation using integrated video conference
application for tele-consultation purposes. Its easy-to-use GUI enables the medical
practitioners to explore the application with the basic level of training (Centre for
Development of Advanced Computing, 2010).
3.6.4.3 E-Sushrut
Healthcare is one of the most important aspects of human life. C-DAC's Sushrut,
a Hospital Information System (HIS) has been developed with the objective of
streamlining the treatment flow of a patient in the hospital, while allowing doctors and
other staff to perform to their peak ability, in an optimized and efficient manner (Centre
for Development of Advanced Computing, 2010).
3.6.4.4 Tejas
Web based Telemedicine software integrated with Hospital Information System
software. It comprises of all essential models including core services, support services
and back office (Centre for Development of Advanced Computing, 2010).
86
3.6.4.5 E-Chavi
As an extension e-Sushrut has another project called e-Chavi, which is an
exclusive image/picture archival system use full for storing the high resolution patient
reports. All these Information Management System products are extremely handy for a
patient who doesn’t need a lot of thought time to access and sort his medical reports (The
blog of Asian hospital & health care management, 2011).
3.7 Communication and Information Infrastructure Including the Internet and Web
In information technology and on the Internet, infrastructure is the
physical hardware used to interconnect computers and users. In some usages,
infrastructure refers to interconnecting hardware and software and not to computers and
other devices that are interconnected. However, to some information technology users,
infrastructure is viewed as everything that supports the flow and processing of
information (Search Data Center, 2011).
3.7.1 Internet
Internet, sometimes called “The Net”, is a very large network of computers. A
more accurate description would be a network of computer networks. Almost all of these
networks use a computer protocol called TCP/IP (Transmission Control Protocol/Internet
Protocol). (Addyman, 1994). Several investigators examined the status and use of
Internet in different types of libraries and application of Internet in various library
functions such as reference work; classification and cataloging; interlibrary loan;
document delivery service; and collection development (Abdoulaye and Majid, 2000;
Adeleke and Olorunsola, 2007; Hundie, 2003; Moore, 2000; Raschke, 1993; Saeed et al.,
2000; Seddiqui, 2003; Trhub, 1999).
3.7.1.1 Search Engines
Search engines are Web sites that maintain an index and short summaries of
billions of pages on the Web, Google being the world's largest. Most search engine sites
are free and paid for by advertising banners, while others charge for the service. Yahoo!
87
was the first search engine to gain worldwide attention, and it initially indexed most of its
content manually, creating a hierarchical directory that was put together by human
observation. It was known then as a "directory" rather than a "search engine." However,
as Web content grew exponentially, it became impossible to index everything manually.
(Your dictionary, 2011)
3.7.1.2 Electronic Mail (E-Mail)
Application software system originally developed for store-and-forward text
messaging over a packet-based computer network. E-mail originated in the mid-1960s for
communications between time-share computer users (Your dictionary, 2011). Electronic
mail which allows users to communicate in fast speed has become a very popular means
of information exchange all over the world. Its use enables both academics and students
to send and receive academic information, exchange research ideas and generally
facilitate cheap and effective communication (Akinseinde and Adomi, 2004).
3.7.1.3 TCP/IP (Transmission Control Protocol / Internet Protocol)
A computer communication protocol is a description of the rules computers must
follow to communicate with each other. TCP/IP is the communication protocol for
communication between computers on the Internet. TCP/IP stands
for Transmission Control Protocol / Internet Protocol. TCP/IP defines how electronic
devices (like computers) should be connected to the Internet, and how data should be
transmitted between them (W3 Schools, 2011).
3.7.2 World Wide Web (WWW)
The World Wide Web (WWW) began as a project at the European Organization
for Nuclear Research (CERN), where Tim Berners-Lee developed a vision of the World
Wide Web.
Tim Berners-Lee - the inventor of the World Wide Web - is now the Director of
the World Wide Web Consortium (W3C) (W3 Schools, 2011). A technical definition of
the World Wide Web is: all the resources and users on the Internet that are using the
88
Hypertext Transfer Protocol (HTTP). A broader definition comes from: "The World
Wide Web is the universe of network-accessible information, an embodiment of human
knowledge." (SearchCRM, 1997)
3.7.2.1 Web Browsers
Quite simply, a browser lets you see information on websites. Currently, browsers
let you see this information in graphical form. The first program that could display
graphical information over the Internet to a user was published in 1993. This first
graphical web browser was called NCSA Mosaic. Prior to its release, the only available
web browser was a text browser named Lynx. In the time since the release of Mosaic,
several other browsers have been created with various features. Internet
Explorer is Microsoft’s browser, Firefox is the Mozilla browser and Apple has released
Safari. (Your dictionary, 2011)
Using Web browsers, libraries have designed and adopted Websites to collect and
preserve information. According to Brower (2004), almost as soon as Mosaic, the world’s
first Web browser, was created in 1993, academic health science libraries (HSLs) began
developing Websites. In 1994 approximately thirteen university medical libraries had
some kind of Web presence in US.
3.7.2.2 Markup Languages
Markup languages are not programming languages, they are sets of markup tags
to describe Web pages.
3.7.2.2.1 HTML (Hyper Text Markup Language)
HTML is a markup language. The purpose of a web browser (like Internet
Explorer or Firefox) is to read HTML documents and display them as web pages. The
browser does not display the HTML tags, but uses the tags to interpret the content of the
page (W3 School, 1999).
89
3.7.2.2.2 SGML (Standard Generalised Markup Language)
SGML is an international standard for markup languages. The basis for HTML
and XML (Web Glossary, 2011).
3.7.2.2.3 XML (EXtensible Markup Language)
XML is a markup language much like HTML. It was designed to carry data, not
to display data. XML tags are not predefined. You must define your own tags. XML is
designed to be self-descriptive. XML is a W3C Recommendation (W3 School, 1999)
3.7.2.3 Web Page
A Web page or web document is a text file coded in HTML, which may also
contain JavaScript code or other commands (Your dictionary, 2011). A web page may
incorporate elements from other websites with suitable markup anchors. Web pages are
accessed and transported with the Hypertext Transfer Protocol (HTTP), which may
optionally employ encryption (HTTP Secure, HTTPS) to provide security and privacy for
the user of the web page content. The user’s application, often a web browser, renders the
page content according to HTML markup instructions onto a
3.7.2.4 Web Site Administration
Web administrators are system architects responsible for the overall design and
implementation of an Internet Web site or intranet. A person responsible for the
implementation of a Web site. Webmasters must be proficient in HTML as well as one or
more scripting and interface languages such as JavaScript and Perl. They may also have
experience with more than one type of Web server (Your dictionary, 2011). Core
responsibilities of the webmaster may include the regulation and management of access
rights of different users of a website, the appearance and setting up website navigation.
Content placement can be part of a webmaster’s responsibilities, while content creation
may not be.
90
3.7.3 Subject Gateways
A simple definition of a subject gateway is "An Internet accessible collection of
descriptions and location details for a range of information, generally available
electronically, organised by subject or discipline, and selected for inclusion based on a
published set of quality criteria." (National Library of Australia, 2007)
3.7.4 LAN/WAN
A LAN (local area network) is a group of computers and network devices
connected together, usually within the same building. By definition, the connections must
be high speed and relatively inexpensive. They function to link computers together and
provide shared access to printers, file servers, and other services. LANs in turn may be
plugged into larger networks, such as larger LANs or wide area networks (WANs),
connecting many computers within an organization to each other and/or to the Internet.
A WAN (wide area network), connects several LANs, and may be limited to an
enterprise (a corporation or an organization) or accessible to the public. The technology is
high speed and relatively expensive. The Internet is an example of a worldwide public
WAN (Indiana University, 2010).
3.7.5 OPAC (Online Public Access Catalog)
OPAC, an acronym for Online Public Access Catalog, is a database composed of
bibliographic records describing the books and other materials owned by a library or
library system, accessible via public terminals or workstations usually concentrated near
the reference desk to make it easy for users to request the assistance of a trained reference
librarian. Most online catalogs are searchable by author, title, subject, and keyword and
allow users to print, download, or export records on an e-mail account (Reitz, 2004).
3.8 Assess, Select and Apply Current and Emerging Information Tools and Create
Information Access and Delivery Solutions
Scientific research is published largely in the form of journal articles, rather than
in books. The Libraries only subscribe to a limited number of journals. A great deal of
information is therefore sought from sources to which the Library does not subscribe
91
(Chambers, 1999). Document Delivery Service (DDS) is the provision
of published or unpublished documents in hard copy, microform, or digital format,
usually for a fixed fee upon request. (Reitz, 2004).
3.8.1 Chemical Abstracts Services Document Detective Service (CASDDS)
Mainly aimed at institutions, but indirectly available to the end-user as it is the
supply source behind SciFinder (also produced by CAS). SciFinder is aimed solely at the
end-user and offers document supply on references retrieved through searches. It is
possible to search DDS itself to a limited extent, to see if a journal is covered by the
service, and for what timeframe. Subject coverage is particularly strong for chemistry
journal articles and patent literature (Chambers, 1999).
3.8.2 Blackwell’s Uncover
Service is available to institutions and to individuals. Users are charged per item
ordered. Provides ``contents'' access to 17,000 journals via a searching interface that
allows retrieval by author, keyword or journal title. Subject coverage is widespread but
there is heavy US bias (Chambers, 1999).
3.8.3. BIOSIS Document Express
Claims to cater for any market, but as payment is by deposit account or invoice
only, this is less likely to attract the individual end-user. BIOSIS Document Express is
not searchable. BIOSIS database coverage offers strong biomedical subject coverage and
includes references to abstracts and book literature not generally found in other
biomedical databases (Chambers, 1999).
3.8.4 ISI Document Solution
Service is available to institutions and individuals. Document solution is not
searchable itself, but similar to CAS and BIOSIS, encourages requesting by use of the ISI
accession number retrieved from searching a related ISI database such as Current
Contents or Science Citation Index. It is possible to request without an accession number,
by inputting standard bibliographic details. Although successful outcome of such a
92
request is not so assured, subject coverage of ISI databases is very wide which gives a
good chance of success on all requests (Chambers, 1999).
3.8.5 British Library Document Supply Centre (BLDSC)
BLDSC now welcomes orders directly from individuals, while continuing to
receive business from many thousands of libraries worldwide. Individual, nonregistered
customers may only order photocopies, which are supplied via the ``Lexicon'' service and
include a variable copyright fee per item. Individuals may not request items for loan.
Individuals dealing directly with BLDSC are also obliged to open a billing account for
payment (Chambers, 1999).
3.9 Competency to Apply Expertise in Databases, Indexing, and Information
Analysis and Synthesis to Improve Information Retrieval and Use in the
Organization
In this part database management systems (DBMS) different indexing methods,
database searching techniques and strategies, some medical and allied health databases,
and data mining and data warehousing are defined.
3.9.1 DBMS (Database Management Systems)
As one of the oldest components associated with computers, the database
management system, or DBMS, is a computer software program that is designed as the
means of managing all databases that are currently installed on a system hard drive or
network. Different types of database management systems exist, with some of them
designed for the oversight and proper control of databases that are configured for specific
purposes. (WiseGEEK, 2003)
3.9.2 Databases Search Strategies
Databases are repositories for storing and retrieving information. In order to find
the information you need, you have to know how to construct an effective search
equation within a database’s parameters. The following searching tricks will help to
93
construct effective searches within not only different databases but many online search
engines, like Google, as well (Saint Mary’s University of Minnesota, 2010).
3.9.2.1 Logical Boolean Combination of Terms
Placing AND between two or more keywords commands the database to retrieve
only results including all of the keywords entered. This narrows the search and increases
the relevancy of the search results. Placing OR between Search terms is the great way to
search for synonyms of words. OR broadens the search, allowing the database to find just
one of the search terms. Placing NOT before a search term excludes that specific term
from the search (Saint Mary’s University of Minnesota, 2010).
3.9.2.2 Truncation of Terms
Another way to expand the search is to truncate it. Truncation locates multiple
forms of a word, including different endings. By adding an asterisk (*) to the end of a
root word, search results will include any applicable variation of the search term (Saint
Mary’s University of Minnesota, 2010).
3.9.2.3 Keyword, Uncontrolled Vocabulary Searching
Keywords are words that describe or relate to the research topic. One should
determine what they think are the best keywords. When searching in a database, one can
choose to look through specific fields—title, author, and abstract etc.—to locate articles
that mention the search term (Saint Mary’s University of Minnesota, 2010).
3.9.2.4 Controlled Vocabulary Searching
Journal articles and books are assigned terms that describe them—controlled
vocabulary—for database organization and retrieval purposes. Controlled vocabulary can
also be called subjects or descriptors. When you search for a resource using the correct
controlled vocabulary terms, you will retrieve relevant, topic-specific results. No need to
worry about weeding out irrelevant (Saint Mary’s University of Minnesota, 2010).
94
3.9.2.5 Pre-Coordinated Vocabulary Searching
Pre-coordinated vocabulary searching refers to the combination of individual
concepts into compound or complex subjects at the point of storage. (Cubbins, et al.,
2007)
3.9.2.6 Post-Coordinated Vocabulary Searching
Post-coordinated vocabulary searching refers to the combination of individual
concepts into compound or complex subjects at the point of retrieval (Cubbins, et al.,
2007).
3.9.2.7 Searching Weighted Terms
A search based on frequencies of the search terms in the documents being
searched. It produces a numerical score for each possible document. A document’s score
depends on the frequency of each search term in that document compared with the overall
frequency of that term in the entire corpus of documents. A common approach
is called tf.idf which stands for term frequency inverse document frequency. The term
frequency means “the more often a term occurs in a document, the more important it is in
describing that document. Inverse document frequency means the more documents a term
appears in, the less important the term is (Howe, 2010).
3.9.2.8 Searching Classified Arrays
Classified arrays are a group of coordinate subjects on the same level of a
hierarchical structure, e.g., oranges, lemons, limes, but not citrus fruit (Cubbins, et al.,
2007).
3.9.3 Medical and Allied Health Databases
This section presents an overview of medical and allied health databases
including Pubmed, Medline, Toxicology Literature Online (TOXLINE), AIDSinfo,
International Pharmaceutical Abstracts, Chemical Abstracts, EMBASE: Excerpta
Medica, Cochrane Library, Pars Medline, Iran Medex, and IndMED.
95
3.9.3.1 PubMed
PubMed, available via the NCBI Entrez retrieval system, was developed by
the National Center for Biotechnology Information (NCBI) at the National Library of
Medicine (NLM), located at the U.S. National Institutes of Health (NIH). Entrez is the
text-based search and retrieval system used at NCBI for services including PubMed,
Nucleotide and Protein Sequences, Protein Structures, Complete Genomes, Taxonomy,
OMIM, and many others. PubMed provides access to citations from biomedical
literature. LinkOut provides access to full-text articles at journal Web sites and other
related Web resources. PubMed also provides access and links to the other Entrez
molecular biology resources (National Center for Biotechnology Information, 2008).
3.9.3.2 MEDLINE
MEDLINE is the NLM's (National Library of Medicine) premier bibliographic
database that contains over 18 million references to journal articles in the life sciences
with a concentration on biomedicine. A distinctive feature of MEDLINE is that the
records are indexed with NLM's Medical Subject Headings (MeSH). The database
contains citations from 1950 to the present, with some older material. New citations that
have been indexed with MeSH terms, publication types, GenBank accession numbers,
and other indexing data are available daily (Tuesday through Saturday) and display with
the tag [PubMed - indexed for MEDLINE] (National Library of Medicine, 2011).
3.9.3.3 Toxicology Literature Online (TOXLINE)
TOXLINE is produced by the National Library of Medicine. The information in
TOXLINE is taken from a number of secondary sources. It contains bibliographic
citations on the toxicological, pharmacological, biochemical and physiological effects of
drugs and other chemicals. It references journal articles, reports, monographs, meeting
abstracts, papers, letters and theses. (The University of San Francisco, 2011)
96
3.9.3.4 AIDSinfo
AIDSinfo is a U.S. Department of Health and Human Services (DHHS) project
that offers the latest federally approved information on HIV/AIDS clinical research,
treatment and prevention, and medical practice guidelines for people living with
HIV/AIDS, their families and friends, health care providers, scientists, and researchers.
(AIDSinfo, 2011).
3.9.3.5 International Pharmaceutical Abstracts
Worldwide, comprehensive bibliographic coverage of pharmaceutical science and
health related literature. Produced in co-operation with the American Society of Health-
System Pharmacists International Pharmaceutical Abstracts includes 30 years of in-depth
indexed reference to the world pharmacy (in the broadest sense) literature; plus, related
health, medical, cosmetic journals, and state pharmacy journals; abstracts of presentations
at major pharmacy meetings are also included. IPA is unique in its coverage-no other
service provides this. The numerous references to alternative and herbal medicine (over
10,000 references from 1970-present) are an example of this unique coverage. (Ovid,
2011)
3.9.3.6 Chemical Abstracts
Chemical Abstracts database includes over 20 million citations to the worldwide
literature of chemistry and its applications from 1967 forward. The database includes
journal articles, patents, reviews, technical reports, monographs, conference and
symposium proceedings, dissertations, books, and journal article preprints. CA SEARCH
corresponds to the bibliographic information and complete indexing found in the print
Chemical Abstracts published by CAS (Chemical Abstracts Service). The controlled
vocabulary CA General Subject Index Headings, related general subject terminology
from the CA Index Guide, and CAS Registry Numbers, each with its modifying phrase
are included. Chemical substances are represented by CAS Registry Numbers, unique
numbers assigned to each specific chemical compound (Dialog, 2009)
97
3.9.3.7 EMBASE: Excerpta Medica Database
A major biomedical and pharmaceutical database well known for its international
scope and timely, in-depth indexing. EMBASE from Elsevier B.V. is a comprehensive
pharmacological and biomedical database renowned for extensive indexing of drug
information from 4,550 journals published in 70 countries (1974-present). It is one of the
most current biomedical databases available. Fully-indexed citations and complete author
abstracts appear on average twenty days after receipt of the journal -- most appear earlier.
Each record contains the full bibliographic citation, indexing terms and codes. More than
80% of the records contain abstracts. The database includes EMTREE, a hierarchically
ordered controlled thesaurus, which contains 46,000 preferred terms and more than
200,000 synonyms. Specialty subsets of EMBASE are also available (Ovid, 2011).
3.9.3.8 Cochrane Library
The Cochrane Library – an online collection of databases that brings together in
one place rigorous and up-to-date research on the effectiveness of healthcare treatments
and interventions, as well as methodology and diagnostic tests. There are over 4,500
Cochrane Reviews currently available in The Cochrane Library. As many as 2,000
protocols for Cochrane Reviews are also available, providing an explicit description of
the research methods and objectives for Cochrane Reviews in progress (The Cochrane
Collaboration, 2011).
3.9.3.9 Pars Medline
With the aim of enabling researchers in the field of health and medicine in Iran to
gain access to research papers in this field, ParsMedline was designed in early 2003 and
became functional in September, the same year. The database comprises more than
79,000 bibliographic descriptions from the years 1978 to 2004. Pars Medline's specialty
fields include medicine, nursing, dentistry, microbiology, virology, special health care
system, nutrition, pharmacology, health, laboratory technology, speech therapy,
audiometry, physiotherapy, and more. More than 95% of the indexed articles have
abstracts and about 33% of them include full text. The Persian journals include both
98
Persian and English abstracts. About 1,000 records are added to the site every month. The
database is accessed approximately 100 times every day (Beiki and Beiki, 2005).
3.9.3.10 IranMedex
IranMedex database is an Iranian database, which includes abstract and full text
of papers issued in Journals printed by medical universities of Iran (Rokni, 2005).
3.9.3.11 IndMED
The purpose of IndMED is to index selected peer reviewed medical journals
published from India. It supplements international indexing services like PubMed. It
covers about 77 journals indexed from 1985 onwards. IndMED has been produced under
an ICMR (Indian Council of Medical Research) funded project - "National Databases of
Indian Medical Journals". Another resource produced under this project is a portal of full
text articles (medIND) of selected Indian medical journals indexed in IndMED or
PubMed. (IndMED, 2011)
3.9.14 Abstracting and Indexing
Abstracting and indexing is a category of database that provides
bibliographic citations and abstracts of the literature of a discipline or subject area, as
distinct from a retrieval service that provides information sources in full-text (Reitz,
2004).
3.9.14.1 KWIC (Keyword in Context)
An acronym for Keyword in Context, a type of permuted index in which
the title of a document (and sometimes the text) is used to illustrate the meaning of a key
word used as an entry. Tagged by hand or extracted from the document algorithmically,
keywords are printed in alphabetical order at a fixed position in a line of fixed length
(usually at the center), so that they appear in a column, with as much of the context as can
be accommodated preceding and following each word. The keywords in the column may
be distinguished typographically to make them easier to read. Keyword and context are
usually coded to identify the document indexed. The KWIC index system accepts an
99
ordered set of lines, each line is an ordered set of words, and each word is an ordered set
of characters. Any line may be "circularly shifted" by repeatedly removing the first word
and appending it at the end of the line. The KWIC index system outputs a listing of all
circular shifts of all lines in alphabetical order (Reitz, 2004).
3.9.14.2 KWOC (Keyword out of Context)
An acronym for Keyword out of Context, a variation on the KWIC (Keyword in
Context) index, in which keywords extracted algorithmically from the title of
a document(and sometimes the text) are printed as headings along the left-hand margin of
the page, with the titles or portions of text containing each keyword indented under the
corresponding heading. A symbol may be substituted for the keyword in the string of
text. Unlike KWAC indexing, this method does not preserve multiword terms andphrases
in the alphanumeric sequence of headings (Reitz, 2004).
3.9.14.3 Rotated and Permuted Keyword Indexing
A type of subject index in which significant words, usually from the titles of the
works indexed, are used as headings. When a string of keywords is rotated, such an index
is said to be permuted (Parker, 2011)
3.9.14.4 Statistical Approaches in Indexing
Statistical systems use programs that extract words or phrases from the text on the
basis of statistical (frequency of the word or phrase within the text) criteria (Compton,
2010).
3.9.14.5 Computer-Assisted Indexing
Machine-aided indexing (sometimes referred to as computer-aided indexing) is
the use of computer software programs to categorize and index unstructured data.
Machine-aided indexing has been used for many years by libraries, data centers and
government agencies to access documents and information through online catalogs and
fee-based information retrieval services, and now this technology is being used to ease
user access to the exploding information available on the Internet (Compton, 2010).
100
3.9.14.6 Controlled Vocabulary Indexing
A controlled vocabulary is a list of terms or other symbols used in indexing.
Terms not belonging to a controlled vocabulary are called free text terms, natural
language terms, and sometimes keywords. The idea of a controlled vocabulary is to
reduce the variability of expressions used to characterize the document being indexed,
e.g. by avoiding synonyms and remove ambiguity (homonyms). Lists of subject
headings and thesauri are typical examples of controlled vocabularies. By principle one is
only allowed to use terms from the controlled vocabulary in the indexing process
(Hjorland, 2007).
3.9.14.7 Pre-Coordinated Indexing
Pre-coordination is the combining of elements into one heading in anticipation of
a search on that compound heading (Wiggins, 2007).
3.9.14.8 Post Coordinated Indexing
Post-coordination is the combining of headings or keywords by a searcher at the
time he/she looks for materials in a catalog (Wiggins, 2007).
3.9.14.9 Use of Weighted Terms
An index term’s weight is based on a function associated with the frequency of
occurrence of the term in the item. Values for the index terms are normalized between
zero and one. The higher the weight, the more the term represents a concept discussed in
the item (Allen, n.d.).
3.9.15 Information Analysis
Information analysis is the ability to discover and quantify patterns in data of any
kind, including numbers, symbols, text, sound and image. The relevant techniques
include statistical and data mining or machine learning methods such as rule induction,
artificial neural networks, genetic algorithms and automated precise systems. (Skills
Framework for the Information Age-SFIA, n.d.)
101
3.9.15.1 Data Warehousing
The term data warehousing comes more from the commercial sector than
academic sources. It comes from the need for organizations to control the proliferation of
digital information ensuring that it is known and recoverable. Its goal is to provide to the
decision makers the critical information to answer future direction questions. Data
warehouses are similar to information storage and retrieval systems in that they both have
a need for search and retrieval of information. But a data warehouse is more focused on
structured data and decision support technologies. In addition to the normal search
process, a complete system provides a flexible set of analytical tools to mine the data.
(Kowalski, p. 21, 1997).
3.9.15.2 Data Mining
Data mining (originally called Knowledge Discovery in Databases-KDD) is a
search process that automatically analyzes data and extract relationships and
dependencies that were not part of the database design. Most of the research focus is on
the statistics, pattern recognition and artificial intelligence algorithms to detect the hidden
relationships of data. (Kowalski, p. 21, 1997).