E ComTech Ref1
E ComTech Ref1
Commerce
Framework, Technologies and
Applications
Fourth Edition
Author’s Profile
Bharat Bhasker is Professor in the information Technology and Systems at
the Indian Institute of Management, Lucknow and former Dean of IIM
Lucknow. He received his B.E. in Electronics & Communications
Engineering from University of Roorkee; M.S. and Ph.D. in Computer
Science from Virginia Polytechnic Institute and State University, USA. He
has worked at Goddard Space Flight Center of NASA, MDL Information
Systems and Sybase, USA, in leading research and research management
positions. Dr Bhasker made research contributions in NASA’s Distributed
Access View Integrated Database (DAVID), Universal Books Management
System (UBMS), NASA’s Data Archival and Distribution Service project
and High Performance Computing and Communications (HPCC) initiatives
at Goddard Space Flight Centre of NASA. He was awarded NASA’s
Research Productivity Award in 1994 in recognition of the research
contributions. He has also served as visiting faculty at University of
Maryland, College Park, University of California, Riverside, University of
Texas, Dallas, Chung-ang University, Seoul, Korea and Essec Business
School, France.
McGraw Hill Education (India) Private Limited
Published by McGraw Hill Education (India) Private Limited
P-24, Green Park Extension, New Delhi 110 016
Electronic Commerce: Framework, Technologies and Applications, 4e
Copyright © 2013 by McGraw Hill Education (India) Private Limited. No
part of this publication may be reproduced or distributed in any form or by
any means, electronic, mechanical, photocopying, recording, or otherwise or
stored in a database or retrieval system without the prior written permission
of the publishers. The program listings (if any) may be entered, stored and
executed in a computer system, but they may not be reproduced for
publication.
This edition can be exported from India only by the publishers,
McGraw Hill Education (India) Private Limited.
Print Edition:
ISBN (13 digit): 978-1-25-902684-3
ISBN (10 digit): 1-25-902684-1
Ebook Edition:
ISBN (13 digit): 978-93-392-1430-2
ISBN (10 digit): 93-392-1430-7
Vice President and Managing Director: Ajay Shukla
Head—Higher Education Publishing and Marketing: Vibha Mahajan
Senior Publishing Manager—B&E/HSSL: Tapas K Maji
Manager (Sponsoring): Surabhi Khare
Assistant Sponsoring Editor: Anirudh Sharan
Senior Production Manager: Manohar Lal
Senior Production Executive: Atul Gupta
Assistant General Manager—Higher Education Marketing: Vijay Sarathi
Assistant Product Manager: Daisy Sachdeva
Junior Product Specialist: Megha Mehra
Senior Graphic Designer (Cover Design): Meenu Raghav
General Manager—Production: Rajender P Ghansela
Manager—Production: Reji Kumar
Information contained in this work has been obtained by McGraw Hill
Education(India), from sources believed to be reliabel. However, neither
McGraw Hill Education (India) nor its authors guarantee the accuracy or
completeness of any information pulished herein, and neither McGraw Hill
Education (India) nor its authors shall be responsible for any errors,
omissions, or damages arising out of use of this information. This work is
published with the understanding that McGraw Hill Education (India) and
its authors are supplying information but are not attempting to render
engineering or other professional services. If such services are required, the
assistance of an appropriate professional should be sought.
Typeset at Tej Composers, WZ 391, Madipur, New Delhi 110 063 and
printed at Nice Printing Press, 3 Rashid Market Extn., Delhi – 110 051.
Cover Printer: SDR Printers
RZCCRRCODZDYY
Dedications
This book is dedicated to my uncle,
Dr Ram Vilas Bajpai,
who taught me that it is important to follow the heart
rather than the crowd.
PREFACE TO THE FOURTH EDITION
This book owes its existence to the course ‘Internet Applications in Business
Management’ offered at IIM Lucknow. During my summer visit to
University of California at Riverside, I was motivated by Dr Satish Tripathi,
Dean of Engineering, to write a book that fills the need for a book that
follows an integrated approach, especially from the Indian context. The book
would not have been possible without the constant support and inputs offered
by Dr Tripathi over these years.
The fructification of this book involves much valued contributions from a
number of persons. I would specially acknowledge the contributions made by
Mr Rajiv Kaka, Mr Satwick Tandon and Ms Kavitha Rao from the PGP batch
of 2000, who compiled the material for the chapters on security, payment
models, and agents in electronic commerce. I would like to further
acknowledge the contributions from Ms Kavitha Rao who wrote the case on
HLL’s Intranet. I also appreciate the discussions and contributions made by
Prof R Srinivasan, IIM Lucknow and Prof Diptesh Ghosh, IIM Ahmadabad
in assisting me in setting up Internet Commerce Research Center (ICRC)
(http: //icrc.iiml.ac.in) at IIM Lucknow in 1999. ICRC has provided us a
platform for focusing our efforts in the emerging field of electronic
commerce through web-based surveys and case developments. Some of the
material developed under ICRC appears in the book as well.
The book has benefitted from the reviews and comments from a set of
anonymous reviewers arranged by McGraw Hill Education (India). I would
like to acknowledge and thank them for helping me in reshaping and
restructuring some of the chapters. Further, I would like to acknowledge the
support, feedback, and encouragement provided by Mr Tapas K Maji, Ms
Surabhi Khare and Mr Anirudh Sharan and the meticulous copyediting and
production management work done by Ms Hema Razdan and Mr Manohar
Lal.
Finally, I would like to express my deepest thanks to my wife Nandita and
daughters Anumeha and Anika for the constant support and encouragement. I
appreciate the sacrifices they had to make and manage the life on their own,
while I was busy completing the project.
BHARAT BHASKER
CONTENTS
Intra-organization Integration
Finally the web can integrate the legacy systems based on mainframes with
other systems across the organization, thus, helping the organization to
expand the information available to decision makers, by integrating existing
systems and by giving them a web face. The richest source of legacy data is
still in mainframe systems, but the easiest data to access is stored in SQL
databases on UNIX, OS/2, or Windows NT servers. By employing web based
tools that use Java applets and or other object oriented Common Gateway
Interface (CGI) libraries, the data stored in existing databases can be
seamlessly integrated.
Intranets result in publication of information inside companies, through
the world wide web, resulting in a paradigm shift in the way in which
information is distributed in an organization. Web based publication and
distribution offers instant, consistent, and correct information to all eligible
users compared to paper based methods. Various departments may benefit
from the intranet in several ways. For example, a basic problem of sales and
marketing departments is delivering up to date reference information to
people distributed over a large geographic area. Salespersons require the right
information at the right moment and the right place to clinch sales.
Through an intranet, salespersons can access the latest information on a
corporation-wide information repository. Product development applications
often centering on project management, with team members updating project
schedules and sharing information about the progress of development or
customer feedback, make an ideal application for an intranet based solution.
Similarly, customer service and support teams can benefit from intranets as it
enables them to share up to date status reports of problems. All the team
members can respond to customer calls, be alerted immediately to any
important changes like special offers or issues, and train online to respond to
customer queries and complaints. Some of the important applications of the
intranet are:
Electronic Sales Information Management A basic problem of sales
and marketing departments is delivering up to date reference
information to people distributed over a geographic area. Salespersons
require the right information at the right moment and the right place to
clinch sales. An intranet helps salespersons in accessing the head office
or the design department with queries from any location provided they
have been armed with an internet enabled tablet, laptop, PC or other
mobile device. The intranet application can be built to provide online
and up to date sales and product information to sales representatives on
the field. As a result, rather than spending time on trying to update
themselves with information they find more time to develop and interact
with clients, leading to more sales opportunities and satisfied clients. In
addition, it also amounts to a great deal of saving in printing and postage
costs on the information that was earlier sent to sales and field offices.
Product Development Product development teams need up to date
information to perform their jobs effectively. Product development
applications often center on project management with team members
updating project schedules and sharing information about the progress of
development or customer feedback. The application greatly assists in the
coordination and communication of design iteration, suggested and
incorporated changes and delivery schedules.
Information Updates The company keeps its employees up to date by
maintaining daily direct downloads of industrial as well as company
news. All the employees stay abreast of the business environment
changes happening around them.
Customer Service and Support The customer support team members
can remain connected and up to date on the status of various reported
problems. It enables them to respond to customer calls and receive
immediate alerts to any important changes like special offers or issues.
Team members can learn from each others experiences in addressing
similar problems through the shared database that maintains logs of
customer problems and solutions. The intranet can be further used for
training them to respond to customer queries and problems online.
ILLUSTRATION 1.8 Hewlett Packard
Hewlett Packard (HP) and Silicon Graphics two major computer systems
and workstations manufacturers deployed intranets in mid-nineties to
improve the intra-organization process efficiency and better information
distribution amongst offices dispersed geographically. HP, an early Silicon
Valley company, has over 25,000 products, including electronic
instrumentation, computer servers and workstations, electronic components,
calculators, and software packages. The company maintains a global
presence through about 600 sales, support, and distribution offices in more
than 100 countries. The intranet deployed by HP runs on close to 2500 web
servers and 170 cache servers to boost the performance of message transfer.
Its intranet handles over 1500 thousand e-mail messages per day and is
accessed by over 100,000 employees of the company through more than
100,000 computers everyday. The company uses this network for a wide
range of activities such as collaborative team work, training, document
management, software distribution, and global electronic communication.
In a company where the corporate culture has always encouraged open
communication among employees, the intranet has truly enhanced the
sharing of information. This has contributed to greater organizational
flexibility, leading to an increase in employee productivity, faster time-to-
market, better customer relations, reduced costs, and the introduction of
more competitive products and services.
Electronic Banking
The increase in penetration of personal computers in home segments has led
to the emergence of several financial management software packages such as
Quicken, Microsoft Money, and Peachtree. Software packages such as
Quicken permit users to organize, interpret, and manage personal finances.
Using Quicken, users record and categorize all financial transaction on a PC.
The user can later use the software to balance the checkbook, summarize
credit card purchases, track stocks and other investments. Personal finance
management through these software packages requires duplication of efforts,
i.e., once by the financial institution and once by the user. Without online
integration with financial institutions to transfer money from his brokerage
account to the money market account, the user sends a paper instruction to
the financial institution, enters it in the personal systems and the bank enters
it in the system to execute the transaction. In addition, the mechanism is also
prone to synchronization problems, forcing users to spend time in discovering
and correcting the anomalies.
With the wide availability and access of the internet, electronic banking
empowers consumers to access their accounts, carry out transactions through
web browsers or web enabled personal software packages, thus, keeping the
two in synchronization as well. Customers can view account details, transfer
funds, pay bills, order checks, and review account history.
ICICI Bank, Citibank, HDFC Bank, and IndusInd Bank have been
offering internet banking services for the past few years.
ILLUSTRATION 1.10 ICICI Bank
ICICI Bank, founded in 1994, has been a pioneer in internet banking in
India. It introduced internet banking in 1997 and has been augmenting the
offerings and services delivery since then. The ICICI initiative was honored
by the Computer Society of Indias (CSI) National Award for best IT usage
in 1998. The Financial Times of UK adjudged the ICICI web site as a
highly commended business site for the years 1997 and 1998. It also
received the coveted cyber-corporate of the year award at the India Internet
World, 1998. The share of internet banking business has been steadily rising
at ICICI Bank. The number of internet customer account grew from 4000 in
March 1999 to 24000 by December 1999. In 2002, the ICICI Bank has
become the second largest bank with assets of Rs1 Trillion and network of
over 1000 ATMs and 500 branches and offices.
ICICI Bank is also an innovator in technological usage for providing
banking services via its branches, ATMs, telephone, personal computer and
the internet. Since April 2000, it started offering Wireless Access Protocol
(WAP) enabled, banking services to mobile customers, through tie ups with
Orange and Airtel Cellular phone service providers. As a result it has
emerged from the shadow of an e-commerce innovator to a technologically
experienced internet bank.
The bank offers convenient access, anytime customer service with the
convenience of 24 X 7 access to accounts through the internet, and
complete control of accounts with the capability to create customized
transaction reports and the facility to make online payments.
ICICI Bank services are based on Infinity from Infosys, India and the
credit card business uses Vision Plus from PaySys, USA. The bank makes a
great effort to protect the security and privacy of transactions, account data,
and personal information. During the account opening process the user
sends the required information using a secure channel. On receiving the
complete set of information, the bank verifies it and then creates a new
account for the customer. When a customer account is created, the bank
assigns a password that is sent to the customers along with an account
verification letter package.
The bank employs a multilayered security model to ensure the
confidentiality of transactions across the internet. At the user end it sets up a
secure session with the ICICI Bank server, using Secure Socket Layer
(SSL) protocol, to provide privacy for the data flowing between the browser
and the bank server. SSL provides a secure channel for data transmission
through its 128 bit encryption capability. The secure channel is utilized for
transfer of information in authentication procedures, providing message
integrity and ensuring that the data cannot be altered during transmission.
The payment gateway for ICICI was set up by Compaq. It uses Compaq
hardware and a QSI Payments Inc. solution for implementing the payment
gateway. The QSI Payments Inc, solution is also used by customers like
HSBC, Hongkong; Merway Bank, BBS Bank, Oslo; Wall-Mart, USA; and
Yappi Credit Bank, Istanbul.
The initial payment gateway solution for ICICI ran on two ProLiant
5500 servers, with having two CPUs each with a SCO UnixWare 7.1
operating system. The NonStop Clusters for UnixWare was deployed to
offer clustering, increased reliability and ensure avoiding single failures.
ICICI was the first financial intermediary to implement an e-commerce
payment gateway within India. ICICI shares the services of the payment
gateway with corporate clients, consumers, merchants and bankers. ICICI
services available under PaySeal™ are used by many B2C electronic
commerce sites to enable the interface of the internet shopper, the web
merchant and the banking systems, in a secured environment to facilitate
online payments. Corporate clients and B2B e-commerce companies also
use ICICI payment gateway e-commerce transactions in a virtual
marketplace.
By exchanging messages using the authentication and encryption
technology of ICICI payment gateway, customers can be assured that they
are actually communicating with the bank, not a third party trying to
intercept the transaction. When a session is encrypted a key or lock icon
appears at the bottom of the browser’s screen. If the key icon appears
broken or the lock does not appear, encryption is not in use and the current
session is not secure.
Electronic Searching
Telephone directories listing personal phone numbers and business phone
numbers play an important role in locating the person or business as the case
may be. The listing of business phone numbers is often organized by business
classifications to assist in locating a business for a particular function. Many
a time phone companies assist by permitting people to ask for information by
description as well. The emergence of the internet and electronic commerce
technologies have been exploited to ease this task by putting the information
a few key strokes away from people connected to the internet. A web browser
can be used for accessing the functionality offered by telephone directories,
by interfacing the directory database with the web (HTTP) server. The
complete functionality offered by a telephone directory service provider can
be offered through a single web interface without any human intervention, all
the time, from all the locations. Companies like Whowhere.com, and
yp.intel.com not only serve the purpose but can additionally provide a lot
more relevant information including travel direction and a map of the
vicinity.
In addition the world wide web has emerged as a vast sea of information.
It contains personal pages, business pages, and general information on almost
each and every topic and subject. Locating relevant information in an ocean
of over 1.3 billion pages can be a Herculean task. Companies like Yahoo
(http://www.yahoo.com), Altavista (http://www.altavista. com), Google
(http://www.google.com), Khoj (http://www.khoj.com), and India123
(http://www.india123.com) have successfully deployed the power of
information retrieval systems and text search engines along with the internet
as a delivery vehicle, through the framework of World Wide Web. These web
applications like Yahoo! Altavista, Google, India123, and Khoj make the task
of searching and locating relevant information easier as well as more difficult
at the same time. Searching based upon concepts, keywords, or subject matter
becomes easier due to availability of powerful search tools. But, searching
may result in a set of thousands of documents, so finding a document
containing the relevant and useful information in the vastly large and ever
increasing web pages can be an arduous task.
Marketing
Traditional marketing practices have relied upon one way communication
due to the nature of the media. Surveys to steer the direction of a company, to
gauge consumer preferences, inclinations and barriers took time to collect,
process, and publish. Traditional marketing faces following major challenges:
Higher Costs The company incurs costs in producing brochures and
product data sheets and in shipping and mailing them to customers.
Supporting consumer queries further require human resources.
Hit Ratio Direct mail, even in targeted market places, suffers from
extremely low response rates.
Time Intensive Marketing tasks are often time constrained, leading to
intense time pressure in organizing the activity. The preparation of an
advertisement or a marketing communication brochure may require
several rounds of revisions, leading to delays in dealing with advertise
agencies and printers. Also, the prepared advertisement may sometimes
have to wait for a long period due to availability of a suitable slot in the
media.
Internet and electronic commerce technologies have been utilized in
mitigating some of these problems. Internet enabled marketing is not a
substitute for traditional marketing, but has emerged as a good augmenting
mechanism. With the interactivity offered by the internet, the marketing
communication need not be a one-way mode anymore. The internet can be
used as media by itself for delivering communication including
advertisements. Several new models have already emerged and have given
rise to a multibillion dollar internet advertising industry. Web sites set up by
various organizations become a ubiquitous medium for marketing
communication. The web page has established itself as a media for banner
advertising in the past few years. Internet advertising offers the following
salient advantages:
Cost Savings Catalogues, brochures, product specifications prepared in
the electronic form and delivered through the internet offer huge savings
in copy editing; printing, packaging and shipping costs, and updating as
and when required. Also, it cuts the time to put the information in the
customer’s hands and up to date information is available to customer’s
worldwide, continuously through the reach of the internet.
Lower Barrier to Entry The size of business, location of business, and
the brick and mortar infrastructure does not matter when you are present
on the internet. The electronic commerce universe is a great leveler. It
offers equal opportunities to one and all by lowering barriers to access
the marketplace.
Interactivity and Information Richness Marketing teams can develop
interactive rich media based brochures, product specifications, and 3-D
views of products and operating scenarios, and place them on the web
site. Analytical buyers can use the information to get enough
information to make an informed decision through interaction with the
site.
Alternate Channel For existing businesses, electronic marketing opens
up a new channel that gives customers the opportunity to browse, collect
information, analyze and then chose the standard product or customize it
to their taste (e.g., color, size, shipping method) and then place the
purchase order. Through interactivity in the customization process, the
customer is more likely to get exactly what they want and the seller is
more likely to clinch the deal.
Electronic marketing offers additional mechanisms and supplements
traditional marketing by providing it a faster access to the global market
space, in a cost efficient manner. In the long term, with an increasing number
of people connected on the internet the electronic market space itself may
grow beyond the traditional market space and will supplement the traditional
marketing strategy making space for the emerging new market space.
Supply Chain Management
The inter-organizational business process that chains the manufacturer,
logistics companies, distributors, suppliers, retailers and customers together
to facilitate order generation, execution, and fulfillment, has evolved over the
past quarter of a century. In addition to product quality, customers deal with
businesses depending upon their ability to execute the handling and delivery
reliably and promptly. Supply chain management deals with three issues:
1. coordinating all the order processing activities that originate at the
customer level, such as the process of order generation, order
acceptance, entry into order processing system, prioritization,
production, and material forecast;
2. material related activities such as scheduling, production, distribution,
fulfillment and delivery; and
3. financial activities such as invoicing, billing, fund transfer, and
accounting.
The process of supply chain management makes a good application
candidate for electronic commerce technologies. It enhances the scope of
supply chain management beyond the efficiency and cost reduction
perspective to growth in revenues, profit margins and improved customer
service. Electronic commerce technologies assist in linking and managing
digitized products, product information, processes, and intercommunication
among organizations. The primary goal of streamlining the product delivery
from the manufacturer to the customer can be better served with digital
communication, sharing of information databases and coordination across a
number of organizations in the ‘chain’. Through the use of internet standards
such as Java and XML, members of a supply chain can pool together
heterogeneous resources and skills for sharing and exchange of information,
to deliver the outcome as one “virtual” organization.
The emergence of virtual organizations is driven by three powerful forces,
viz., the globalization of the economy; restructuring of industry due to
emerging economic realities and WTO; and the emergence of electronic
commerce, driven by internet technology for a new mode of interaction
between manufacturers, suppliers, distributors, and customers. In the face of
global competition, the inefficiency, high production costs, and outmoded
products are taking a pounding. Products will be manufactured where it is
cheapest and most efficient to make them. The monolithic vertical-
manufacturing model is already facing an immense pressure and adjusting to
it. Today, even market leaders such as IBM, HP and Apple who made most
of their own components and assembled almost everything in-house have
resorted to outsourcing, complete with Original Equipment Manufacture
(OEMs), ECMs (electronic contract manufacturers), EMSs (electronic
manufacturing-service providers), independent designers, suppliers, and
distributors. The virtual corporation derives a competitive edge by creating
networks of specialized companies. In the network each company specializes
in a certain sub process or subassembly in which it is the best. Electronic
commerce and communications technologies interconnect these processes
along with the information exchange standards and protocols to provide the
shape of the virtual corporation.
Electronic Trading
Electronic trading, in short is a mechanism that utilizes the power of
electronics and communication media, such as the internet, to bring together
geographically dispersed buyers and sellers on a virtual common trading
platform. The common platform offers aggregated information to all
participants in a fair manner. The platform facilitates access to aggregate
information, order booking, and fulfillment.
In the context of stock markets, e-trading means buying and selling equity
online through electronic means. In practical terms, it is accomplished
through registered brokers such as ICICIdirect, Etrade, Fidelilty and Charles
Schwab, to name a few. The buyers and sellers registered for electronic
trading, rather than relying on phone conversations to track and collect
information followed by faxed orders to buy or sell, can use the do-it-yourself
paradigm. Investors can access their accounts with the broker by logging on
to the network. The investors are provided with up to date market information
and may decide to enter a buy or sell order online. Orders in the electronic
trading environment are executed directly without any manual interventions.
The entered order is executed and fulfilled based upon investor-defined
constraints. Electronic trading in stocks is accomplished through brokers.
Brokers in electronic stock trading provide execution only services in
contrast to full service brokers and advisory brokerage services. Full service
brokers offer complete investment service—the money is handed over to the
brokerage account and the broker manages the money. It is the broker who
decides when and what stocks to buy and sell on behalf of the client and
charges him for the services. In the advisory service account, the broker
offers advice on what to buy, sell or hold in your account but the final
decision rests with client. Finally, the executions only service brokers simply
do what the client tells them. As a result, they also offer services at the
cheapest rates. These brokers are often referred to as discount brokers due to
lower service charges. In the electronic trading environment, all the market
information is available to the investor, who is probably the best judge of his
money, investments and risks. As described earlier, electronic (online)
brokers are execution only brokers, who accept orders on the system through
network or even touch tone phones. Trading online offers the following
advantages over traditional means.
Cost Electronic trading is based on accepting an electronic order over
the network, entered through digital computing devices. Brokers need
reliable servers, that are much cheaper than manning a bank of
telephones and fax stations, for accepting and then entering those orders.
As a result, the cost of transaction is comparatively cheaper in electronic
brokerage. The broker passes on some of the savings in transaction costs
to the investors/customers.
Accessibility An investor has access to the account 24 hours a day and 7
days a week. They can access the account, check account balances,
execution status, and analyze account performance at a time of their
convenience. Investors can enter orders, even when the markets are
closed, for later fulfillment.
No Queues With online trading, the issue of waiting on phone lines,
especially when the customer is eager to know the status or make a
trade, is happily resolved. In phone-based trading it may not be
economical for brokers to have lines to meet the peak demand with no
waiting. In online trading the broker can maintain enough bandwidth
and server computing power to handle the peak load.
The electronic trading model has been widely adopted in the stock
trading/brokerage markets. Etrade (http://www.etrade.com) began offering
web based brokerage services in mid 1990’s with aggressive advertising
campaigns and became a brokerage house to reckon with. Leading discount
brokers like Charles Schwab, Quick and Really, and Fidelity followed suit.
Even Merrill Lynch, which had steadfastly held on to the non-discount
brokerage model for nearly 85 years, had to succumb to market trends. In
1999, Merrill Lynch launched web based trading with a competitive price
structure. Although stock trading remains the major application of electronic
trading, it has been successfully applied in the area of trading of chemicals,
gases, and electrical equipment, among others.
SUMMARY
Concept of Electronic Commerce and Benefits: Electronic commerce is
evolving the processes involved in commerce by introducing electronic
means. In the process, it improves upon traditional commerce by making it
efficient and reducing transaction friction. Elements of the market and how
they benefit from electronic commerce are discussed.
Impact of Electronic Commerce: The transformations brought about by
electronic commerce have been impacting market structure, businesses,
consumers, and society at large. The impacts on these entities due to the
transition towards electronic commerce, along with the inherent risk and
measures that need to be taken to mitigate the risk are discussed.
Electronic Commerce Classifications: Electronic commerce involves a
transaction between two parties. The type of entities involved in the
transaction influence the mode and nature of information sharing, payment,
and delivery mechanism and also at times the type of electronic network, and
who can access it. Electronic commerce has been classified into B2B, B2C,
C2B, C2C, and intra-organizational commerce, based on the entities
involved.
Electronic Commerce Applications: In several areas electronic commerce
applications have been successfully used for the past few years. Some of the
areas where it has been thriving are auctions, banking, searching, education
and learning, marketing, supply chain management, and stock trading.
REVIEW QUESTIONS
1. What is electronic commerce and how does it differ from traditional
commerce?
2. What is “friction” in a transaction? Identify sources of friction in
electronic commerce transactions.
3. Define the elements of a market and describe how electronic commerce
influences each of these elements.
4. What is the likely impact of electronic commerce on economic
structures like the industry, consumers and society?
5. Categorize electronic commerce transactions based on the entities
involved.
6. What is supplier-centric B2B electronic commerce?
7. What are the roles of each entity in intermediary-centric B2B electronic
commerce?
8. How can an existing business take advantage of Business-to-Consumer
(B2C) electronic commerce?
9. List the electronic commerce applications described. Identify any two
additional common applications of electronic commerce.
10. What is B2E electronic commerce?
11. What is intra-organizational electronic commerce and what are its
potential benefits?
Thes biggest happening of the just past decade has been the emergence of the
new network driven digital economy. Internet, was born out of the command,
communication, control (C3) initiative of the Defense Advance Research
Projects Agency of the US Government. It was further advanced for its
ability to provide unlimited shareability of information, resources, and a
distributed computing environment, by the academic and research
laboratories. During this decade, the internet made a transition from being a
bastion of non-commercial purity to being the driving force behind electronic
commerce. This transition has had a profound impact. Not only, it has
reshaped the business paradigm, but also society at large the way we
communicate, conduct business, acquire knowledge, and even the way we
play and entertain.
As of 2002, globally, more than 60,000 newer people are joining the
internet economy everyday. These internet economy entrants are utilizing it
for retrieving stock information, managing financial portfolios online, paying
bills, buying books, music, groceries, bidding at auctions for goods,
entertainment, training courses, online competitive exams and sharing online
video archives, music, photographs, or voices with friends, and family.
To top it all, for accomplishing much of it, they are not dependent on the
desktops alone. Wireless devices, palmtops, pagers, and mobile phones have
joined the network to extend the outreach of the network economy. The
mobile devices consortium already boasts of over a million users of the
newly arrived WAP, and WAP enabled devices to access the internet, and
derive the privileges of a networked economy. It is growing at a pace that
will surpass the total desktop internet users, in few years.
The year, 1999 stands out as a landmark year for a boom in network
economy, inspired by the early successes of Yahoo!, Altavista, Amazon,
Ebay, Infospace, CommerceOne, Hotmail Indiaworld, and Rediff, hordes of
others with angel’s, or their own resources, joined the gold rush. During that
year, Dr. Koop, Boo.com, Chipshot.com, ToySmart.com, and closer home,
Autoindia.com, Jaldi.com, Mantraonline.com, Indya.com, Fabmart.com, and
many others joined the race. Many a new entrants with newly acquired
capabilities also put up web sites with some content thinking users will come
rushing in to frequent their businesses.
Unfortunately, business reality has begun to set in. Many of these
businesses have shut their doors, others, are hobbling along. Even the high
growth star Amazon.com has rarely rung in profits, despite being in operation
for over five years. Why are profits non–existent or at best rare in internet
businesses? The answer, as expected, may be that euphoria cannot substitute
for a good business model.
SUMMARY
This chapter introduces and discusses the role of business models in
electronic commerce. There have been a plethora of business models that
have been used for offering commerce over the internet.
The chapter provides a taxonomic survey of business models that have
been used by various businesses operating in electronic commerce
environment. In this chapter, the business models have been categorized on
two dimensions, viz., information content based versus transaction based and
transplanted to internet versus native to internet.
REVIEW QUESTIONS
1. What do you understand by a business model?
2. Describe the taxonomy of the business models used in this chapter. Can
you come up with an alternate taxonomy to classify electronic
commerce business models?
3. Define and differentiate between an infomediary and a metamediary.
4. What is a affiliate model? Provide two examples of electronic commerce
businesses that use this model.
5. What is the electronic store model? What are the major impediments
faced by the model in less developed countries?
6. What are the major advantages of the manufacturer model? Describe
how the model reduces market friction and costs through a value chain
analysis.
REFERENCES AND RECOMMENDED READINGS
1. Berry, J. “A potent new tool for selling: Database marketing,” Business
Week 338 (September 5, 1994): 56–62
2. Bayers, C. “The inner Bezos”. Wired (, March, 1999).
http://www.wired.com/wired/archive/7.03/bezos.html.
3. Clemons, E. and M. Row, “Sustaining IT advantage: The role of
structural differences”, MIS Quarterly 15, no 3 (September, 1991): 275–
292.
4. Cronin, M. J. Doing business on the Internet: How the electronic
highway is transforming American companies, New York: Van
Nostrand Reinhold, 1994.
5. Horwitt, E. “Casting a wider net”. ComputerWorld, (27 July, 1998)
http://www.computerworld.com/home/Emmerce.nsf/All/980727casting.
6. Hagel III, J., and J. F. Rayport, “The new Infomediaries”, The McKinsey
Quarterly, (Number, 1997).
7. Joseph Pine II., B. Mass customization, the new frontier in business
competition, Harvard Business School Press, 1993.
8. Koning, J., M. Occello, N. Ferrand, Y. Demazeau, F. Van Aeken, and C.
Baejis, “A multi–agent approach for mediation support on the net”, 1st
International Workshop on Decentralized Intelligent and Multi–Agent
Systems, Krakow, Poland (November 1995).
9. Martinez, P. “Model made “e”: What business are you in?” Center for
IBM e–Business Innovations
(http://www.ibm.com/services/innovations).
10. McKenna, R “Real–Time Marketing”, Harvard Business Review (July,
1995) 87–95.
11. Nissen, M. E. “Commerce Model and the Intelligent Hub”,
CommerceNet CALS Working Group Presentation. (November 1995).
12. OsterWalder, A. and Y. Pigneur, “An e–business model ontology for
modeling e-business,” Proceedings of the 15th Bled Electronic
Commerce Conference, Bled, Slovenia, June 2002.
13. Rappa, M. “Business Models on the Web, Digital Enterprise,”
http://www.digitalenterprise.org/models/model.html.
14. Resnick, P., J. Zeckhauser, and C. Avery, Roles for Electronic Brokers,
Edited by G. W. Brock, Toward a Competitive Telecommunication
Industry: Selected Papers from the 1994 Telecommunications Policy
Research Conference: Mahwah, NJ: Lawrence Erlbaum Associates.
289–304. http://www.sloan.mit.edu/CCS/ccswp179.html.
15. Sarkar, M. B. “Intermediaries and Cybermediaries: A continuing role for
mediating players in the electronic marketplace,” JCMC 1, No. 3
(December 1995).
Learning Objectives
This chapter covers the following topics:
1. Introduction to the conventional purchasing process
2. What is electronic data Interchange
3. Building blocks of EDI systems
4. Value added networks
5. Benefits of EDI systems
WHAT IS EDI?
Electronic Data Interchange (EDI) is the exchange of business documents
between any two trading partners in a structured, machine-readable form. It
can be used to electronically transmit documents such as purchase-orders,
invoices, shipping bills, receiving advices, and other standard business
correspondence between trading partners. EDI can also be used in
exchanging financial information and payments in electronic form. The
Electronic Fund Transfer (EFT) systems used by financial institutions are a
prime example of the application of EDI in the banking and financial sector.
EDI should not be viewed as simply a way of replacing paper documents and
traditional methods of transmission such as mail, phone, or in-person delivery
with electronic transmission. Rather, it should be seen not as an ‘end’, but as
a means to streamline procedures and improve efficiency and productivity.
EDI covers wide and varied application areas and, depending upon the
perspective, has been defined in several ways. According to the Data
Interchange Standards Association.
“Electronic Data Interchange (EDI) is the computer-to-computer exchange
of business data in standard formats. In EDI, information is organized
according to a specified format set by both parties, allowing a “hands-off”
computer transaction that requires no human intervention or rekeying on
either end. All information contained in an EDI transaction set is, for the
most part, the same as on a conventionally printed document.”
The Webopedia says that,
“Electronic data interchange, is the transfer of data between different
companies using networks, such as the Internet. As more and more
companies get connected to the Internet, EDI is becoming increasingly
important as an easy mechanism for companies to buy, sell, and trade
information. ANSI has approved a set of EDI standards known as the X12
standards.”
According to the EDI University, a training provider in EDI,
“EDI stands for Electronic Data Interchange, a method of transporting all
types of information, such as purchase orders, invoices, payments and even
graphics, to another party electronically. EDI technology was introduced by
Value Added Networks (VANs), in the 1970’s, as an alternative to modem
banks, and essentially replaces paper-based communications with electronic
equivalents. Since EDI is based on a standard developed by the American
National Standards Institute (ANSI), everyone can use it, enabling all
businesses to share a common language.”
The National Institute of Standards and Technology says that,
“EDI is the computer-to-computer interchange of strictly formatted
messages that represent documents other than monetary instruments. EDI
implies a sequence of messages between two parties, either of whom may
serve as originator or recipient. The formatted data representing the
documents may be transmitted from originator to recipient via
telecommunications or physically transported on electronic storage media.”
According to the Electronic Commerce Technical Assistance Group,
“Electronic Data Interchange (EDI) is the computer-to-computer exchange
of business data in standard formats. In EDI, information is organized
according to a specified format set by both parties, allowing a “hands off”
computer transaction that requires no human intervention or re-keying on
either end. The information contained in an EDI transaction set is, for the
most part, the same as on a conventionally printed document.”
The two key features that run through all the definitions narrated above
include the electronic exchange of information, and standard formats or
business forms. The electronic exchange of information requires the presence
of direct or indirect interconnection between the involved partners. The
typical business forms used in EDI: include schedules, purchase orders,
acknowledgements, delivery related documentations, receipt notes, invoices,
remittance requests, payments through electronic fund transfer, bills of
lading, manifests and reconciliations and many other forms depending upon
the application area. These documents have to follow a standard format. The
standardization of format helps in exchanging these documents between
trading partners who may have heterogeneous computing environments.
2. Detail refers to line items that describe the actual business transaction. In
case of a purchase order, it may contain item number, description,
quantity ordered, and price information. In the sample order form shown
below, the detail information refers to the two line items (resistors and
switches) shown below.
BENEFITS OF EDI
1. Reduces Lead Time
In the EDI environment, the exchange of documents among trading partners
happens electronically through interconnected computers. The process of
transferring the documents/information is instantaneous, offering weeks of
time savings compared to the traditional environment that used postal/courier
based exchange of printed documents. Also, the direct electronic transfer of
documents between inter-organizational systems eliminates the chances of
error due to re-entry of data printed on paper from one system to another
system. As it streamlines the information flow, the cycle time is reduced
drastically. In the EDI environment, order-processing, shipping of goods, and
invoice-preparation and transmission can all be done within a matter of a few
hours compared to the days/weeks it takes in a non-EDI environment.
2. Improves Coordination with Suppliers
Traditional trading environments are often burdened with the problem of
mismatched invoices, un-matching terms in quotations and purchase orders,
missing invoices even after the bill for payment is received and many similar
inter-business problems. On careful examination, it will be evident that much
of these problems are caused either by delays in the transmission of printed
documents, loss of documents in transition, or due to errors in the
transcription of the printed information into the electronic form. The
instantaneous transfer of business documents over the network in electronic
form and confirmation of the same addresses the first problem, thereby
making nearly impossible for documents to arrive in wrong sequence. Also,
since the documents are received in electronic form, the need to re-enter the
same data is not there and, as a result, transcription errors are totally
eliminated.
3. Reduces Redundancy
As all the documents exchanged between trading partners are stored in an
electronic mailbox, documents can be accessed, retrieved, and examined at
any point of time. Either trading partner can access, examine, and make a
copy of the document from the electronic box instantly. Contrast it with the
non-EDI system; it may take hours, or even days, to locate and retrieve a
printed business document from the past. Many a time, trading partners file
copies of the same document at multiple places. The EDI environment
eliminates the need for multiple copies and reduces redundancy without
compromising the accessibility and retrieval of old documents.
4. Expands the Market Reach
Most large manufacturers like General Motors deal with EDI-enabled
suppliers only. In the process of streamlining the purchase process they often
institute a value-added network. By being a part of their value added
network, many opportunities open up for supplying the material to some
other larger suppliers who are also a part of the network. Also, with the
growth of electronic commerce and further integration of EDI with electronic
commerce, the creation of an electronic marketplace by large manufacturers
who buy supplies from many large and small suppliers, has become a reality.
By, participating in this large market place you are likely to pick many orders
from other suppliers who are a part of the market/place/network. The General
Electric initiated Trade Process Network (tpn.com) is a prime example of
such a marketplace.
5. Increases Revenue and Sales
Many large organisations use EDI and trade with other EDI-enabled
suppliers. The efficiency brought about by EDI reduces the total transaction
friction by eliminating paperwork and related errors that ensue. It also leads
to quicker settlement of accounts. The reduced transaction friction saves
money and the supplier is in a better position to offer the items at cheaper
costs, leading to improved revenue realisations and sales.
APPLICATIONS OF EDI
The ability to exchange business documents electronically has been found to
facilitate coordination between the partners, reduce the lead-time and thus
reduce inventory. Although, large manufacturing and transportation
companies were the early birds who recognized the advantages, any of the
other industry segments also stand to benefit from electronic document
exchange. The health care, and financial sectors and cross-border trade
facilitated through electronic document exchanges including customs services
—have been some other sectors that adopted and derived the returns from
EDI.
SUMMARY
The paper-based processes deployed in purchase and supply chain
management were cumbersome and time consuming. The processes incurred
significant delay due to the duplication and re-entry of information from one
computer system to another. Electronic data interchange evolved in order to
streamline purchase processes and reduce the duplication of effort due to
paper-based document exchange. In this chapter, we studied the definition of
electronic data interchange and the architectural blocks of electronic data
interchange. EDI systems consist of following layers:
1. Applications/Conversion Layer which defines the functionality of the
actual business application.
2. Standard Formats Layer, which defines the EDI document standards
used by the system. The widely adopted standards are EDIFACT and
ANSI X12.
3. Data Transport Layer, which concerns itself with the protocols that are
deployed for transporting an electronic document from one system to
another system. Protocols such as X.435, email, and FTP, are often used
for transportation purposes.
4. Interconnection Layer, which Concerns itself with the basic connectivity
mechanism needed for transporting electronic documents between the
geographically distributed computers. The Internet, I-way, wide area
networks and Dial-up connection protocols are some of example
protocols deployed in this layer.
REVIEW QUESTIONS
1. What is electronic data interchange?
2. Describe the paper-based ordering process and how EDI alters the
process?
3. Describe the impact of EDI on the supply chain management of a
manufacturing plant.
4. What is the value added network and what are the salient features of a
value added network provider?
5. What are the basic building blocks of an EDI system?
6. What are the advantages of using the common format approach in EDI
systems?
7. What do you understand by, the ANSI ASC X12 standard?
8. What is a transaction set in the ANSI ASC X12 standard?
9. What are benefits of using EDI?
____________________________
1This case has been prepared from secondary source as a basis for class
discussion rather than to illustrate effective of ineffective handling of a
situation.
Learning Objectives
This Chapter covers the following topics:
1. What is the architectural framework of electronic commerce
2. Elements of the Electronic Commerce framework
(a) Network Infrastructure
(b) Information and Distribution Technology
(c) Networked Multimedia Content Publishing Technology
(d) Security and Encryption
(e) Payment Services
(f) Business Services Infrastructure
(g) Public Policy and Legal Infrastructure
SUMMARY
This chapter introduces architectural elements and the framework of
electronic commerce. The framework of electronic commerce requires
technological, business service, and public policy infrastructure.
The technological aspects require a robust, reliable network access,
secure; reliable and portable information distribution; easy to use
information content creation; multimedia publication technology and the
technology to ensure security, privacy, integrity, and authenticated
access to the information content.
The business service infrastructure requires applications for locating and
identifying businesses and the means to carry out a safe and secure
transaction including online payments. The safety and security of
transactions is based on third party trust based assurance for
authentication, privacy, integrity and non-repudiation of a transaction.
The technology can provide the basis for a secure transaction but the
trust environment requires public key infrastructure.
The trust environment has its basis in the public policy and legal
framework. Establishing electronic commerce related laws and
recognized certification authorities provides the legal framework for
electronic commerce.
REVIEW QUESTIONS
1. What are the basic architectural elements of electronic commerce?
2. What is the role of online payment systems in electronic commerce?
3. What are the essential technologies for ensuring security in an electronic
commerce environment?
4. What is meant by business service infrastructure? Compare the business
service infrastructure requirements of traditional and electronic
commerce.
5. What is the role of the certification authority in the electronic commerce
framework?
6. What are the requirements for the creation of a trust environment in
electronic commerce?
INTERNET
The ARPANET protocol, after adoption of TCP/IP, was capable of
interconnecting and communicating across multiple networks. With the
popularity of ARPANET and the associated benefits that emanated to the
academic and scientist community, the number of networks and hosts grew
exponentially. In 1984, National Science Foundation (NSF) of USA
established a backbone connecting six supercomputer centers and around
twenty regional networks that provided connectivity to university campuses.
Adoption of the TCP/IP reference model made it easier to interconnect the
ARPANET, NSFNet, Space Physics Analysis Network (SPAN) of NASA,
High Energy Physics Network (HEPNet), European Academic and Research
Network (EARN), and BITNET. The early backbone of the internet was
formed by the ARPANET backbone and that is why many times confusion
exists between ARPANET and internet.
Today, the internet is characterized by the TCP/IP Reference Model, the
unique addressing scheme, called IP Address, and the Domain Naming
System that makes it possible to uniquely address every host connected on
the internet. A machine is said to be on the internet, if it has an IP address,
runs TCP/IP software and can exchange IP packets from all other machines
on the internet.
All the machines on one network have the same network identifier,
irrespective of the class of the network. The routers interconnect various
networks and switch traffic packets between networks.
Message Preparation and Framing
The IP layer is operated on the top of existing networks, each one with their
own data link layer and associated addressing scheme. The underlying
networks themselves are not aware of the IP addressing. These different
networks are interconnected together through the IP addressing mechanism.
Internet protocol utilizes the existing data link layer of networks by mapping
the data link layer addresses with the IP addresses, encapsulating the
transport (upper) layer message into IP packets and then creating data link
frames in the underlying network format. The original message, encapsulated
as the IP packet and finally framed in the physical networks format, travels
smoothly on the existing network. The upper layers running TCP or UDP
may try sending messages larger than the frame sizes permitted by the
underlying network. The IP layer fragments these messages into smaller
packets so that they can be framed within the size limits of the underlying
networks. On the receiving end the IP layer is responsible for reassembling
these fragments into original packets, prior to delivering it back to the upper
layers. It is this flexibility of IP, to package, fragment, frame, reassemble and
map IP addresses to carrier network addresses, that makes it possible to
interconnect many different networks.
The data link layer frames an IP packet as payload or data. The IP layer
puts in enough information for the data link layer to carry out framing by
collecting and passing all the information along with the IP packet to the data
link layer, so that it can use its regular framing module to generate the frame.
For example, if the IP layer was operating over ethernet, the ethernet will
require a 14-byte header and 4-byte trailer consisting of a cyclic redundancy
code. The header consists of a 6-byte ethernet source, 6-byte ethernet
destination address and a 2-byte type field. The IP layer sends a packet to the
ethernet framing module along with the ethernet address of the source and
destination with the field value implying that the payload data is an IP packet.
The ethernet frame creation module uses these parameters to set the header
fields, places the IP packet in the data field, computes the checksum and
transmits it on the broadcast channel. The IP layer require the ethernet
address of the destination machine, when operating over ethenet, even if it is
familiar with its IP address. The IP address space can be set by the user,
while the Ethernet address remains fixed with the network interface.
Similarly, any physical network whose data link layers are being utilized by
the IP layer for packet exchange has its own address space. The issue that
requires to be addressed is a mechanism through which the IP layer can
dynamically map the IP address to the physical address of the interface. The
task, in the broadcast based physical networks, is accomplished by an internet
support protocol, called Address Resolution Protocol (ARP).
Address Resolution Protocol
The address resolution protocol provides the mechanism for determining the
data link layer address of any IP address in a broadcast based network. If two
devices connected on a local area network want to communicate with each
other at the application level, using TCP/IP, then the applications may set up
a TCP connection for exchange of messages. The TCP messages injected into
IP layer travel on the underlying local area network. The underlying network
in this case has its own data link address. Thus, the IP packet has to be
framed in the local area network frame format, using the data link addresses
of that layer for the delivery.
The IP layer maps the IP address to the data link layer address, using the
ARP. The protocol uses a special request packet. The packet contains the
ARP request code, data link layer (DLL) type, network type, the IP address
and the DLL address of the sender, and target IP address of the machine
whose DLL address is desired. The ARP packet is framed in the data link
layer’s format. In the case of ethernet, the frame header contains the ethernet
address of the sender as the source address and the broadcast address (all 1’s)
as the destination address. The frame is broadcast on the local area network.
Each interface on the local area network receives and processes the request.
The machine that owns the IP address specified in the target IP address field
of the ARP packet, frames a reply to the source DLL address by filling the
target data link layer field and sending it. The IP address mapping request is
broadcast to all the machines connected on the network, but the reply is
marked to the sender of the ARP request. All such translations are cached at
each machine interface to improve efficiency. To address the problem of out
of sync caches, anytime a new machine comes up on the network or an IP
address is changes, a new ARP packet containing the IP address and
corresponding ethernet address is broadcast, causing all caches to be updated
with the latest information.
Transport Layer
The objective of the transport layer in the TCP/IP model is to offer efficient
service for carrying out communication between hosts on the internet. It uses
the internet layers IP service for exchanging information between any two
internet hosts and offers the applications the services of establishing
connection oriented communication or the connectionless exchange of
information. The transport layer of the TCP/IP model supports two protocols
— Transmission Control Protocol (TCP), for providing a reliable, connection
oriented byte stream service; and User Datagram Protocol (UDP), for
providing connectionless, unreliable but faster service. Both of these
protocols are built on the Internet Protocol, which is a connectionless
unreliable protocol.
Transmission Control Protocol (TCP)
The TCP supports a reliable delivery of a byte stream between two end
points, over an unreliable network. In this protocol, two entities trying to
communicate with each other establish a connection. The connection is
established by creating communication end points, also known as sockets.
The socket address consists of two components—the IP address and a 16-bit
number, called port. A port is a transport layer service access point. The
connection is established between the two sockets of the peer machines,
using the service primitives of the TCP. The TCP connection is point-to-point
and full duplex, ensuring that the traffic can move in both directions
simultaneously. Once the two machines have established a connection
through sockets, the byte stream can be transmitted from one end point to
another end point.
The connections in TCP support a byte stream rather than the message
stream. In the message stream, if one entity writes four messages that are 256
bytes long, the receiver will receive four messages. In the byte stream the
four blocks of 256 bytes sent may be received as four blocks of 256 bytes, 2
blocks of 512 bytes, or 1 block of 1024 bytes. The receiver is in no position
to detect the message or the packet boundary. Instead, it receives continuous
sequence of bytes that can be read. The TCP connections behave very much
like UNIX pipes as far as byte stream is concerned.
The data sent by the upper layer (applications) is formatted as a TCP
packet. The protocol may buffer the data till it reaches an adequately efficient
size or may push the data immediately by preparing a TCP packet and
handing it over to the IP layer. The TCP packet is shown in Fig. 5.17.
REVIEW QUESTIONS
1. Describe the characteristics of networks based on broadcast
transmission.
2. What do you understand by network topology?
3. What are the various transmission media used in local area networks?
4. What is CSMA/CD protocol?
5. Describe the IEEE 802.3 standard and its importance.
6. What is a Wide Area Network?
7. Describe the TCP/IP reference model.
8. What is an IP Address? Describe the classes of IP Addresses and
reasons for dividing it in classes.
9. What is a domain name?
10. Describe the domain name system and how it manages the name space.
11. Define the role and purpose of NAP and PNAP in the internet industry
structure.
The internet offers infrastructure for constructing tools that can exchange
information at the application level. Application tool designers and builders
can utilize the internet layer level services, such as socket interface, a
application programming interface (API) that shields the developer from the
intricacies hidden under the network layers. Many of applications developed
over the years have been widely adopted for information exchange and
distribution purposes. These applications also some times referred to as the
standard internet applications, have welldefined and accepted protocols. The
adoption of standardized protocol, for these applications, has further added
toward their acceptance and adoption. Multiple vendors/groups can provide
interchangeable clients and servers for these applications. File Transfer,
Remote Terminal, Electronic Mail, News Groups, and the World Wide Web
are some of the widely accepted internet applications. Each of these
applications follows the client-server model with a standard protocol. File
transfer application is based on File Transfer Protocol (FTP), electronic mail
is based of Simple Message Transfer Protocol (SMTP), news groups are
based on Network News Transfer Protocol (NNTP), and the world wide web
is based on Hypertext Transfer Protocol (HTTP).
In the following sections we briefly describe some of these protocols,
servers and clients.
$ ftp ftp.cdrom.com
Connected to wcarchive.cdrom.com
220—wcarchive.cdrom.com FTP server (Version DG-4.1.73
983302105)ready.
Name (ftp.cdrom.com:bhasker): anonymous
331—Guest login ok, send your e–mail address as password.
Password:
230—Welcome to ftp.cdrom.com, a service of Digital River, Inc.
230—There are currently 496 users out of a possible 3000.
230—
…
230—Guest login ok, access restrictions apply.
Remote system type is UNIX.
Using binary mode to transfer files.
ftp> ls
200 PORT command successful.
150 Opening ASCII mode data connection for ‘file list’.
Total 12
-rw-r—r-- 1 root wheel 480 May 1 16:59 RATECARD.TXT
-rw-rw-r-- 1 root wheel 696 Nov 19 1997 README
-rw-r—r-- 1 root wheel 3344 Sep 1 2000 UPLOADS.TXT
drwxrwxr-x 2 root wheel 512 Oct 5 1998 archives–info
drwxr-xr-x 2 root wheel 512 May 2 1999 etc
drwxrwxr-x 2 root wheel 2048 Jun 26 19:55 pub
226—Transfer complete.
ftp> get UPLOADS.TXT
Local: UPLOADS.TXT
Remote: UPLOADS.TXT
200 PORT command successful.
150—Opening BINARY mode data connection for ‘UPLOADS.TXT’
3344 bytes).
226—Transfer complete.
3344 bytes received in 0.459 secs (7.1 Kbytes/sec)
ftp> bye
221—Goodbye!
ELECTRONIC MAIL
Electronic Mail (e-mail) is an internet application that offers the ability to
exchange messages among users on remote computers. E-mail is the most
widely used application, in fact for many people it is the mainstay
application, rarely do they use other applications. The e-mail application built
upon the TCP stream offers the reliable and instant delivery of messages in a
user’s mailbox. An e-mail system is concerned with the ability to compose
messages, move messages from the originator’s site to the recipient’s site,
report the delivery status to originators, browse messages by the recipients
and finally the dispose off messages. A typical architecture of the e-mail
system (Fig. 6.2) consists of two components to accomplish the functionality
—a user interface program and the message transfer server. The user
interface, also often called mail reader, is a program that offers users an
interface to compose a new message, read a message, reply to senders and
delete or file the message. The user interface program (mail reader) provides
three of the five functions, i.e. composing, browsing, and disposition. There
are a variety of mail readers available. Some of these are built on a character
based interface, driven by the keyboard input, like mail and pine, while others
offer a Graphical User Interface (GUI), that is menu and icon driven and
accepts inputs from the mouse and keyboard. Message Transfer Agent
(MTA) programs accomplish the function of transferring the message to the
destination. These programs communicate with each other using a standard
protocol. A user agent composes a message and informs the message transfer
agent of its delivery, by placing it in the appropriate queue. The composed
message contains the destination mailbox address. The message transfer
agent connects to the other message transfer agent running on the machine
specified in the destination address of the composed message and delivers it
through the standard message transfer protocol. In the internet environment
the Simple Message Transfer Protocol (SMTP) has been widely adopted and
message transfer agents using the protocol are often referred to as SMTP
servers.
In the above sample, the header fields To:, CC:, BCC:, From:, Sender:,
Received:, and Return Path: are used by the message transfer agents. The
fields with ‘X-’ prefix are user-defined fields, used by the Netscape user
agent, and the remaining fields are used for recipients and user agents. The
message body follows the header section. In the text only RFC 822 format,
the body is a free flowing text and users are free to format it the way they
desire. In the extended RFC 822 format the content-type and related fields
add structure to the message. The multimedia information containing non-
textual data is encoded in base64 or quotable print formats, prior to handing it
over to a message transfer agent. The sender, receiver and other addresses
used in the internet environment have acquired a standard format that is
based on the IP address and domain name system. All computers on the
internet have a unique IP address. The domain name system maps a domain
name to an IP address. Thus, the mail addresses used have the
[email protected] format. The domain name (suffixed to @) through the
resolver DNS determines a unique machine and the username identifies a
unique mailbox located on the machine.
Message Transfer
Message transfer agents are responsible for delivering the message to the
destination machine. In the Internet environment, the SMTP is widely used
by message transfer agents. Simple Mail Transfer Protocol (SMTP) is an
ASCII based protocol. In a typical message transfer between two SMTP
daemons, the sender makes a TCP based connection to the daemon running at
port 25 of the machine specified in address field of the header. On successful
establishment of connection, the message is transferred to the destination
daemon using SMTP. A sample session of the protocol conversation is shown
in Box 6.3.
Box 6.3 A Sample Session of SMTP
R: 220 mit.gov.in
S: HELO ganga.iiml.ac.in
R: 250 mit.gov.in says hello to ganga.iiml.ac.in
S: MAIL FROM: [email protected]
R: 250 sender ok
S: RCPT TO: [email protected]
R: 250 recipient ok
S: DATA R: 354 Send mail; end with “.” on a line by itself
S: From: [email protected]
S: To: [email protected]
S: Message Id: <[email protected]>
S: Date: Thu, 16 Aug 2001 20:56:11 + 0530
S: Reply To: [email protected]
S: Organization: Indian Institute of Management Lucknow
S: X-Mailer: Mozilla 4.75 [en] (Win98; U)
S: X-Accept Language: en
S: MIME -Version: 1.0
S: Subject: Meeting Notice
S: Content-Type: text/plain; charset=”iso-8859-1”
S: The first meeting of the Information Technology planning group will
be held
S: at 3:00 P.M. today. Meeting will be held in Conference room CR-108.
S:.
R: 250 message accepted
S: quit
R: 221 mit.gov.in closing connection
The SMTP protocol is defined in RFC 821. The message transfer follows
the envelope and content model. The envelope is constructed from the
“From:” and “To:” fields of the message format. In a typical session between
two SMTP daemons, the receiving daemon on accepting a connection request
from the sender responds by sending a welcome message. The sender
daemon responds with the ‘HELO’ command and informs it about its own
domain. After the handshake phase, the address on the envelope is used by
the sending daemon to establish the data transfer to the right user on the
receiving side. The sending daemon communicates, to the receiver, the
protocol packet containing a ‘From’ address, followed by the recipients’
addresses one at a time. The receiving daemon responds to each of the
protocol packets, either with an “Okay”, or with a specific error message. The
error responses may arise due to various reasons, a common one being the
non-existence of a user mailbox on the receiver side, to whom the mail is
addressed.
Applications of Electronic Mail
In addition to personal communication, electronic mail systems with MIME
capability can be used for distributing the multimedia information. The
electronic mail system has been utilized to provide file transfer facility;
sending remote commands to be processed at the recipient’s machine and
dispatching the processed results; or manage information directories; send fax
through e-mail and facilitate discussion groups. These applications of e-mail
require specialized servers at the receiver’s end. One generic application that
offers information /file management and delivery services is called Mail
Server.
A mail server accepts all the incoming messages destined for a specific
userid and processes the body of message as a list of commands. Typically,
the subject line is left blank and the mail server ignores it.
The mail server running at the machine sends back the files available at
the site. After locating a useful file, the user may send another mail with the
message body containing ‘file <filename>’ and will receive the file by e-
mail. Here is subset of commands that can be commonly sent to servers for
processing.
% mail [email protected]
subject:
file /ls-lR
Help
Document-by-name/send name [name, …]
File path [path, …]
Person name, organization [country]
Whois name
E-mail servers have also been used for offering FTP functionality over e-
mail. In this case all the mail arriving at the specially created account is
picked up by the specially designed server. It is a useful service for users who
do not have FTP access from their own machines. The publicly available
FTPMAIL software can be set up on a machine with complete FTP access.
The FTPMAIL server reads the mail addressed to the FTP e-mail account
and, accepts messages with blank subject line and body containing FTP
commands. Users can send regular FTP commands in the body of the
message, starting with the “connect” command, followed by “Get”, “Chdir”
and other valid FTP commands and, finally ending with a “Quit” command.
WHAT IS HTTP?
Hypertext Transfer protocol is set of rules that world wide web clients and
servers use to communicate over the network. It is a connectionless protocol,
meaning that browsers and servers do not establish a permanent connection.
A client opens a connection and submits a request message to a server. The
server on receiving a message, processes and responds to it and closes the
connection. It is also a stateless protocol, implying that the server does not
maintain any information on the state of the process. Thus, the server treats
each request/message independent of any previous requests/messages. The
protocol is based on the request/response model.
The client, usually a web browser, submits a request to a web server. The
server reads the incoming protocol packet, processes it and sends the
response. The content type is built as part of the protocol’s response packet.
The browser has to be aware of the type of multimedia content delivered to it
as a response. The content types used in the protocol are a subset of the
standard MIME types. As stated earlier the browser connects to the server
machine, specified by domain name/IP address, at the specified or standard
port. On making a successful connection, the browser submits an HTTP
request. A typical HTTP session between the client and server is depicted in
Fig. 6.4. The session consists of two phases, the first phase consists of the
client’s request submission, while the second phase consists of the servers
response. The client submission, depicted in three steps, involves opening a
connection, sending the request and header information.
Step 1: HTTP packets can be transmitted only after the client has established
a connection with the server. In this step the browser parses the URL for
identifying the domain name. It uses the services of Domain Name Server
(DNS) to resolve the name into an IP address. Using the services offered by
the TCP layer, it opens a connection to the IP address, at a standard web or
URL specified port. On the successful opening of a connection, the browser
starts the HTTP session.
Fig. 6.4 Typical Interaction in an HTTP Session
Step 2: The browser submits HTTP packets containing the request command,
to the connected server. The common HTTP request commands are “get”,
“post”, and “head”. The request in HTTP is made up of three components,
viz., the command method, resource identifier and the protocol version
number. An example of the “GET” command is as follows:
GET /index.html http/1.0
The method describes the type of request and determines the response at
the server end. The second component is a resource identifier, such as the
name of a file to be retrieved. Parsing the URL and stripping out the protocol
name, domain name, and port number (if present), derives the resource
identifier from the URL. The last component of the request specifies the
version number of protocol being used. For the URL
http://icrc.iiml.ac.in/index.htm, the browser, after establishing a connection to
the domain name icrc.iiml.ac.in at port 80, would submit the following
request command.
GET /index.htm http/1.0
If the URL entered in the browser window did not include a document
name then a default document name is retrieved. For example, for the URL
http://icrc.iiml.ac.in, the request command would be as follows.
GET /http/1.0
In case of an interactive session, that uses forms for submitting the data to
be processed by the common gateway interface (CGI) mechanism of the
HTTP server, the request line also contains data as a part of the resource
identifier. The details about the CGI mechanism will be discussed later.
The ‘head method’ syntax is identical to that of the ‘get method’. It also
works in the same fashion as the get method, except that the requested
document is never transferred to the browser. The server processes the
request in head request method, and it sends only the header information to
the browser. Usually, it is used for testing purposes. Most link checker
programs, that ensure that a site contains all the existing and valid links,
utilize the head method.
Finally, the ‘post method’ is devised as an alternative mechanism for
submitting the form data entered at the browser end, to the server for
processing. Unlike the ‘get method’ that appends the form data to the
resource identifier, the post method sends the data as a part of the header
information. When a server receives the post command, it knows that the data
will be arriving after the header information, along with the length and type
of the data.
Step 3: In this step the browser submits the header information to the server.
The header information includes the browser identity, its capability to handle
various types of content, and the referring URL. The header information
follows a standard format of header name and the value pair, separated by the
colon (:) sign. The following example shows the header information
transmitted by a browser.
GET / HTTP/1.0
User-Agent: Mozilla/4.75
Referer : http://icrc.iiml.ac.in/
Accept: image/gig, image/jpeg, image/png, */*
Accept-Language: en
Accept-Charset: utf-8, iso-8859-1
The header information is read and processed by the server and is made
available at the server end as environment variables. For example, referer
information is available as the HTTP-Referer environment variable, accept as
HTTP-Accept, and so on. In case of the ‘post method’ the browser as part of
the header information also submits the form-data, content-type and content-
length. The following example shows the headers for the ‘post method’.
POST / myprog.cgi HTTP/1.0
User-Agent: Mozilla/4.75
Referer: http://icrc.iiml.ac.in/
Accept: image/gig, image/jpeg, image/png, */*
Accept-Language: en
Accept-Charset: utf-8, iso-8859-1
Content-type: application/x-www-form-urlencoded
Content-length: 27
----- Carriage Return -------
username=myuserid&name=G+I+JOE
As can be seen the post command, few new headers have been added. The
content-type header available at the server end, as the Content_Type, informs
the user about the MIME type of the arriving data. The content length
available on the server side, as Content_Length, informs the user about the
length, in bytes of the attached content. The content, i.e., data itself is
transmitted as the last part of the header section, separated from the headers
by a new line.
Step 4: On receiving the client request and header information, the server
processes the request and sends the response to the client. If the request was
processed and can be delivered, the server sends an OK response. Some
common errors that it may send as responses include forbidden document,
‘not found’, ‘internal server error’, ‘or’ ‘unauthorized access’. The format of
the response sent by the server includes the response code and the protocol
version. The protocol version informs the client about the kind of syntax used
in responses. Examples of server responses are as follows:
HTTP/1.0 200 OK
HTTP/1.0 404 Not Found
HTTP/1.0 401 Unauthorized
HTTP/1.0 403 Forbidden
The first component of the response informs the client about the protocol
version number used by the server for sending the response, as the syntax of
the response may vary between versions. The second component is the actual
response, consisting of the response code and the message. The clients use
the code part for interpreting the response and acting accordingly. The
message part is displayed to users. In most web servers these messages can
be customized as well. On receiving the response code of “200 OK” the
browser understands that the request was processed successfully and
proceeds to receive the data that it had requested.
Step 5: Prior to sending the requested data, the server sends information
about the data, such as the type of content and length of content as well as
information about the server itself, as part of the response phase. The
response headers sent by servers are also used, at times, for accomplishing
authentication and setting up cookies. The response header information
follows the same syntax as request headers. The following example shows
typical response header information.
Date: Tue, 04 Sept 2001, 10:40:05 GMT
Server: Apache/2.1.2
Last-Modified: Sun, 02 Sept 2001, 08:05:10 GMT
Content-Length: 8455
Content-Type: text/html
The above header information informs the browser of the date and time at
which the server response was sent and the name and version of the server
software. It also informs the browser of document-related information. The
Last-Modified date tells the user when the requested document was last
updated. The last two headers tell the browser about the length of the
requested documents, in bytes and the type of content. In this case the
browser readies for receiving a 8455 bytes long text document of the html
subtype. The html subtype indicates, to the browser, that the document needs
to be parsed, interpreted, and rendered for HTML tags. On the other hand
content-type of text/plain would have been displayed by the browser as it is.
Step 6: The server, after sending the last response header information, sends
a blank line to indicate the completion of header portion the response and to
mark the beginning of the response data. The server sends the response data
to the browser in the format indicated in the content-type response header.
Step 7: The web server, on completing the data transmission, is done with
responding to the client request. At this stage, it would ordinarily close the
TCP connection. However, an HTML document may contain online images
and embedded objects that are required for rendering it on the browser
screen. Although, the browser can submit a request for retrieving each of
these objects, by opening a new connection to the same server, the approach
incurs heavy overheads of opening and closing connections. Network
bandwidth and server efficiency can be improved by keeping the connection
active for subsequent requests. The browser can accomplish this by including
the following request header, in the client request headers, discussed in Step
3.
Connection: keep-alive
In this case the server keeps the TCP connection open even after the
response data has been sent. The browser uses the same connection for the
subsequent request.
SUMMARY
Internet infrastructure lays the foundation for applications that access
and manipulate distributed and remote information. The major and most
widely used network applications are file transfer, Telnet, electronic
mail, and world wide web.
File transfer applications rely on the File Transfer Protocol.
Electronic mail systems use the RFC 821 and RFC 822 specifications.
The specification defines the Simple Message Transfer Protocol
(SMTP). It uses ASCII headers to define the message properties.
The World Wide Web, evolved in late eighties and early nineties, is
based on Hypertext Transfer Protocol. It is a system that hyper links
geographically distributed multimedia documents. Web documents are
pages written in the Hypertext Markup Language (HTML).
World wide web (WWW or web) systems use the client-server
architecture.
Web servers manage the HTML document and handle client requests.
The client and servers interact with each other using the Hypertext
Transfer Protocol (HTTP). Web clients use Uniform Resource Locators
(URLs) for identifying the documents in the Internet universe.
NCSA and CERN were the two early public domain implementations of
the web server and Mosaic was the first Graphical User Interface (GUI)
based client available freely. Since then several commercial browsers
and servers for the world wide web have been available from companies
like Netscape and Microsoft.
Apache software foundation has been constructing and distributing a
NCSA based web server under the public domain software policy.
Apache is a cooperative movement supported by several of volunteers.
REVIEW QUESTIONS
1. Describe the importance of a protocol.
2. Briefly describe the purpose of file transfer protocol and list five
important commands.
3. Assume that the SMTP server is running and accessible to you through
Telnet on icrc.iiml.ac.in. Describe the session log for sending a message
“Happy New Year” to [email protected] [email protected].
4. Describe salient features of Hypertext Transfer Protocol.
5. Define a 7-step interaction between a HTTP client and a server as
described in this chapter.
6. In HTTP (Web) servers of the NCSA lineage, what are the files used for
configuring the web server?
7. What is the purpose of defining the DocumentRoot in the configuration
file of the NCSA lineage web server?
INFORMATION PUBLISHING
The large part of the growth of the world wide web can be attributed to its
ability to integrate a variety of information, seamlessly, from distributed
servers. In the process of integration, the web addresses several issues and
offers following advantages:
1. Platform Transparency Access to the web is through a piece of software
called the browser. Regardless of whether the browser is running on the
Windows, X-Windows, or Macintosh platform, it offers the same interface.
The web is not limited to any single platform or machine. The data residing
in a variety of server platforms is available to users, through the same look-
and-feel interface.
2. Distribution Transparency The web is a distributed information system.
The information, stored at a variety of geographically dispersed server
platforms, is available to the web users on a single interface window. A page
displayed on a browser screen may contain text coming from an IBM server
in New York, an image from Windows NT servers located in Delhi and a
background audio clip from a Linux server in Lucknow. The distributed
nature of the web enables it to successfully provide so much information,
stored in thousands of servers located across the globe.
3. Information Type Transparency The web offers seamless integration of
multiple types of information content. text, graphics, sound, video and
various other data formats can be integrated and displayed uniformly through
the browser interface. It can integrate a variety of information content, stored
on distant servers, through the Hypertext mechanism. The concept of
Hypertext really means that instead of sticking to reading text in a linear,
rigid structure, the important terms can be made rich by adding/linking the
explanation to it. Any time you click on the rich term, the linked explanation
shows up. Some commonly used examples of the Hypertext system are
HyperCard on Macintosh, Help on MS-Windows and Answerbook on Sun
Microsystems. The web not only integrates and handles text, but also a
variety of media. In a true sense it is a hypermedia system.
4. Interactive Information browsing on the web is based on selecting and
clicking on links. Clicking on links retrieves and offers additional
information on-screen. A simple interaction on the web can lead one to a
maze of information. In addition to the simple interactivity, the web also
supports forms with input windows, radio buttons, options lists and
checkboxes for submitting the data. Web servers can collect the input
information from users, through the form mechanism, add it to a database,
update the database, or provide customized information, depending upon the
inputs.
5. Dynamic The information retrieved by browsers is stored in a site and
offered through a web server. At any point of time, if the information is
updated at the server site, the latest version is available to anyone browsing it.
Unlike published documents or books where every new version/edition to be
distributed physically. Web publishing does not incur any cost of reproducing
copies. Anyone accessing the publication reads the latest version (except in
case of cached documents) on their browser screen.
6. Graphical and Navigational The capability of web to integrate and
display graphics, text and other multimedia formats, in color on a same page,
is probably the reason for its gaining popularity over such a short span of
line. Prior to the web, the information on the internet was accessible through
command and menu based interface. One could download text or even
graphics and then had to invoke appropriate tools to browse the content, that
too on separate windows on the screen. The web has made it possible to
browse multimedia information on the same page. The hyperlinking
mechanism has also reduced the task of navigating through the information to
point-and-click. A user can jump from pages stored on one server to another
server just by clicking on links.
WEB BROWSERS
Web users access information from web servers, through a client program
called browser. Broadly speaking, a browser is responsible for the following
tasks. The first task is to accept a URL and retrieve the document specified
by the URL. In the process of retrieving the document it parses the URL into
its components, i.e., the protocol, domain name, port number, and document
name. The client program connects to the web server specified by the domain
name and port number and the subsequent conversation in HTTP retrieves the
document. Since browsers are capable of accepting URLs with other
protocols such as FTP, they are conversant with other protocols and retrieval
mechanisms as well. However, most often browsers deal with HTTP and as a
result retrieve documents written in the HyperText Markup Language
(HTML). A HTML document includes in its structure, text, hyperlinks to
other documents, images, and multimedia information. The second task of the
browser is to interpret a HTML document code, format it accordingly and
finally render the document on the screen. In the process, it may have to
manage the rendering of various image formats, multimedia information, and
links to other documents as well.
The earliest, most popular, client program for the web was developed by
the National Center for Supercomputing Applications (NCSA), at the
University of Illinois at Urbana-champaign. This browser, called Mosaic,
offered graphical user interface and displayed the text and graphics in full
color on the same page. Mosaic was made available on Macintosh, MS-
Windows and X-Windows platforms. The browser for all the three platforms
mentioned above can be downloaded from the site http://www.ncsa.uiuc.edu/.
For example, the MS-Windows version can be downloaded by accessing the
URL http://www.ncsa.uiuc.edu/SDG/Software/mosaic-
w/releaseinfo/download.html.
Members of the Mosaic development team founded a company called
Netscape Communication Corporation. The Netscape browser was made
available, free of cost for personal use, on the above three platforms. The
Netscape browser became the most popular browser within a year of its
release. It was later acquired by America Online and still remains one of two
major web browsers in the internet space. Information regarding
downloading the Netscape browser can be found at the site
http://www.netscape.com. It can be downloaded from the site or any of the
listed mirror sites, closest to your location.
The Internet Explorer of Microsoft is another popular browser, based on
the Mosaic source code. Like its counterpart Netscape, the Internet Explorer
has been defining, creating and offering up-to-date features and is available
for MS-Windows, Macintosh and Unix environments. For the MS-Windows
environment, it is usually available as a part of pre-loaded software. The MS-
Windows version of the software can be downloaded either directly from the
site http://www.microsoft.com/windows/ie/downloads/ie6/default.htm or
from its mirror sites.
For the text only internet connections, a browser called Lynx that lets you
access the basic information and supports hyperlinks, was developed by the
University of Kansas. The Lynx browser offers navigation through the use of
arrow keys. The browser software can be downloaded from the site
ftp://ftp2.cc.ukans.edu/pub/lynx.
The second type of text formatting tags define the physical formatting
style. The impact of these tags is consistent across browsers. Some sample
tags are as follows:
The body of the document shown in example 1 can be formatted with the
tags described here. Example 2 shows the document with the formatting tags.
The resulting rendering of the document in a browser is shown in Fig. 7.2.
Example 2
<html>
<head>
<title> My First Document</title>
</head>
<body>
<h1> Examples of Formatting Tags</h1>
Using <EM> Notepad</EM> or any <B>text editor </B>, you can type
the document as it is. Save the file as <I> exm2.htm </I>
<br>
Invoke the Browser and open the file <strike> myhtml.htm </strike> <I>
exm2.htm </I> to see the output on your screen.
</body>
</html
Fig. 7.2 Browser View of the Example 2 Document
Block Structuring Tags
As stated earlier, browsers ignore carriage return, line break, white spaces
and tabs. The document is made up of blocks. The blocks contain a specific
kind of text and may have some common associated properties. For example,
a document contains one or more paragraphs, each paragraph starts on a new
line and has an associated alignment property. The text in a paragraph is
processed and formatted by the browser. On the other hand, at times we may
be interested in formatting the text and ensuring that it is displayed in exactly
in the same format on the browser as well. To achieve this HTML provides a
<pre> tag that ensures that the formatted text within the tag pair is displayed
as it is in a browser. Block structuring tags are used for creating the blocks
such as paragraphs, addresses, blockquotes and preformatted text. Following
are some of the example tags:
Below, Example 3 illustrates the usage of some of these tags. The browser
view of the document is shown in Fig. 7.3.
Example 3:
<HTML><HEAD>
<TITLE>Preformatted Text</TITLE>
</HEAD>
<BODY>
<p>
The following text appears in user-formatted form:</p>
<pre>
This is an example of preformatted text.
Large States in India.
1. UP 2. MP 3. Maharashtra
North Eastern States:
1. Assam 2. Mizoram 3. Arunachal Pradesh
4. Manipur 5. Nagaland 6. Meghalaya
7. Tripura
</pre>
<p> This is an Example of Blockquote tag:</p>
<Blockquote>
Uttaranchal, Jharkhand and Chhattisgarh are the three
most recent states of India.
</Blockquote>
</BODY>
</HTML
Fig. 7.3 Browser View of the Example 3 Document
List Tags
List elements are used for organizing part of document in an annotated
listing. The lists themselves can be numbered (ordered) or bulleted
(unordered). There are a wide variety of lists, supported by HTML. These
include definition list, directory list and menu list. The ordered and unordered
lists are two primary types of the lists. The ordered list is marked by tags
<OL> and </OL>. Similarly, the unordered list is marked by tag pair <UL>
and </UL>. Within the tag pair, the beginning of each list item is marked by
the <LI> tag. The markup tag <LI> does not require a matching end tag as
the appearance of <LI>, </OL>, or </UL> as the case may be implicitly
marks the end of list item description. There are other types of lists that are
used for describing the menu items or defining terms. The markup tags for
these lists are <DL> and <Menu> respectively. Example 4 illustrates the
usage of ordered, unordered and definition list tags and Fig. 7.4 depicts the
rendering of the example code by a browser.
Example 4:
<html>
<head>
<title> My Document: Ordered and Unordered Lists</title>
</head>
<body>
Fruits <br>
<OL>
<LI> Apples
<LI> Oranges
<LI> Bananas
<LI> Pears
</OL>
Flowers <br>
<UL>
<LI>Lotus
<LI>Rose
<LI>Marigold
<LI>Jasmine
<LI>Sunflower
</UL>
Computer Devices <br>
<DL>
<DT>CPU<DD> Central Processing Unit
<DT>ALU<DD> Arithmetic Logic Unit
</DL>
</body>
</html>
Fig. 7.4 Browser View of the Example 4 Document
Image Tag
Images, in the HTML document, can be included using the <IMG> tag. This
tag, like many other HTML tags, has several attributes associated with it.
These attributes, in case of the <img> tag, can be used defining alignment,
width, height and the name of the image source file. Below, Example 5
shows the usage of <img> tag in a HTML document. The Fig. 7.5 depicts the
rendering of the code by the internet browser.
Example 5:
<html>
<head>
<title> Example of Image in a Document</title>
</head>
<body>
Here is an example of the including a image in the document.<br>
<img width=40% src=”anumeha.gif”>The images in gif89a formats can
be animated.
</body>
</html
#!/bin/sh
echo Content-type: text/plain
echo
echo Output of the Who command on icrc
/bin/who
The output of the above script, on execution (if only two users are logged
on at the time), will be as follows:
In case of the get method, the data is made available to the CGI
program/script through the query_string environment variable. In the case of
the post method, the information is made available to the CGI program/script
as the standard input (stdin) stream. The server is not obligated to send the
end-of-file (EOF) marker to the input stream. Thus, the program/script has to
rely on the content _length environment variable. The executed CGI
program/script specified by the action attribute in the form tag extracts the
values entered by the user. In order to extract the values the program/script
examines the environment variable request_method. If the examined value of
the variable contains get it extracts the argument string from the query_string
environment variable. In case of the post method, the script will read the
number of character specified by the content_length environment variable,
from the standard input (stdin). The following CGI script written in shell
script on Unix echoes some of the environmental variables set by the server,
received by program/script and discussed above.
#!/bin/sh
echo Content-type: text/plain
echo
echo CGI/1.0 test script report:
echo SERVER_SOFTWARE = $SERVER_SOFTWARE
echo SERVER_NAME = $SERVER_NAME
echo GATEWAY_INTERFACE = $GATEWAY_INTERFACE
echo SERVER_PROTOCOL = $SERVER_PROTOCOL
echo SERVER_PORT = $SERVER_PORT
echo REQUEST_METHOD = $REQUEST_METHOD
echo HTTP_ACCEPT = “$HTTP_ACCEPT”
echo PATH_INFO = “$PATH_INFO”
echo SCRIPT_NAME = “$SCRIPT_NAME”
echo QUERY_STRING = “$QUERY_STRING”
echo REMOTE_HOST = $REMOTE_HOST
echo REMOTE_ADDR = $REMOTE_ADDR
echo REMOTE_USER = $REMOTE_USER
echo AUTH_TYPE = $AUTH_TYPE
echo CONTENT_TYPE = $CONTENT_TYPE
echo CONTENT_LENGTH = $CONTENT_LENGTH
For the form shown in Example 8 and the values assumed, the output of
the program is shown as follows:
Fig. 7.11 Browser View of the Output of the Script Program
The script written in Unix shell displays the content of selected
environment variables. The shell script writes the output as per the
specifications of the CGI, to the standard output (stdout) file. The first line of
the program is a directive that specifies the location of the shell that needs to
be executed. The second line writes the content-type message to the output
file, while the next line writes a blank line. The blank line acts a separator
between the output content and the CGI header portion. The rest of the lines
echo the content of selected environment variables. The server receives the
script generated output, formats it as a HTTP reply packet by adding the
required header information and creates the envelope. The received reply
envelope is displayed on the user screen (Fig. 7.11) by the browser.
The following example script (written in Perl) extracts the user submitted
input from any form and displays the content back to the user in the encoded
form. Let us assume that the form shown in Fig. 7.10 is used with get method
and the action field refers to the following CGI script.
#!/usr/local/bin/perl
$request = $ENV{‘REQUEST_METHOD’};
if ($request == “get”)
$query = $ENV{‘QUERY_STRING’};
else {
$q_length = $ENV{‘CONTENT_LENGTH’};
read(STDIN,$query, $q_length);
}
print “content-type: text/plain\n”;
print “\n”;
print “Echo of the user input in Encoded form:\n\n”;
print “---Begin Content ---”;
print $query
print “---End Content---”;
Content-type: text/plain
<CR>
Echo of the user input in Encoded form:
---Begin Content ---
U sername=bharat+bhasker&agegroup=midage&reading=yes&riding=
yes
---End Content---
The output is displayed, back to the user, through the browser screen as
shown in Fig. 7.12. The above script example demonstrates the mechanism of
extracting form submitted user input values, in the encoded format. The
following script (written in Perl) decodes and parses the query string to
extract the values and print them.
#!/usr/local/bin/perl
$request = $ENV{‘REQUEST_METHOD’};
if ($request == “get”)
$query = $ENV{‘QUERY_STRING’};
else {
$q_length = $ENV{‘CONTENT_LENGTH’};
read(STDIN,$query, $q_length);
}
print “content-type: text/plain\n”;
print “\n”;
print “Echo of the user input in decoded form( name=value):\n\n”;
print “---Begin Content ---”;
$query =~ s/\+/ /g;
# label 1
$query =~ d/%([\dA-Fa-f][\dA-fa-f])/pack(“C”,hex($1))/eg; #label 2
@pairs= split(/&/, $query);
#label 3
for each $pair (@pairs) {
#label 4
($label,$value) = split(/=/, $pair);
print $label,”=”,$value,”\n”;
}
print “---End Content---”
Fig. 7.12 Browser View of the CGI Program Output
The above script (written in Perl) extracts the query string and then in step
1 substitutes all the ‘+’ characters with spaces. In the line labelled 2, it
translates back the hex encoded special characters to the original form. The
resulting string ($query) is parsed to extract the “name=value” pairs by the
split operator. The ‘&’ character is used as the field pair separator in the URL
encoded encryption type. Thus, splitting the ‘&’ and assigning the values to
an array of named pairs in step 3, all the name=value pairs become elements
of the array. The name and appropriate program variables can then be
assigned values. In the loop step labelled 4, each pair out of the @pairs
(array) is selected and further split on the ‘=’ character and assigned to $label
and $value variables. The above script for the same form inputs, described in
previous examples, send the following output to the web server:
Content-type: text/plain
Echo of the user input in decoded form(name=value):
---Begin Content ---
Username=bharat bhasker
agegroup=midage
reading=yes
riding=yes
---End Content---
The web server sends the output by appropriately placing all the headers
through HTTP. On receiving the output, the browser renders the output on
screen as shown in Fig. 7.13.
Various programming/scripting languages commonly used for CGI
purposes have built-in library functions for extracting values. For example,
the CGI object library of Perl contains functions (param) that can retrieve the
value of any given field in the form by its name attribute. CGI programs can
process values manipulate them, or store them in databases in the usual
fashion. Or, the input can be used for retrieving relevant and related
information and a response can be constructed dynamically from the database
and presented to the user.
mydoc.shtml
<HTML>
<HEAD>
<TITLE>Server Side Includes IllustrationM</TITLE>
</HEAD>
<BODY>
The following document displays Server Side Includes Directives:<br>
Document Name: <!--#echo var=”DOCUMENT_NAME”--><br>
Date: <!--#echo var=”DATE_LOCAL”--><br>
This file was last modified on: <!--#echo var=”LAST_MODIFIED”- ->
<br>
Also the size of mydoc.shtml <!--#fsize file=”mydoc.shtml”-- ><br>
</BODY>
</HTML>
The server, on parsing the file, interprets all the SSI directives and
replaces them by their respective values. The echo directives are replaced by
the value of the environmental variable. The fsize directive, usage fsize
<filename>, is replaced by the size of the file provided as argument.
<filename> is replaced by the size. All the environmental variable listed
earlier with the CGI can be displayed. In addition, the following
environmental variables can be used with echo directives:
Variable Description
DOCUMENT_NAME The file name of the current document.
DOCUMENT_URL The URL of the current document.
QUERY_STRING_UNESCAPED The query string submitted with all shell
characters, escaped with the backslash
character.
DATE_LOCAL The date as per the server’s local time
zone.
DATE_GMT The date as per GMT format.
LAST_MODIFIED The date and time when this file was
last modified.
We have already seen two of the server directives, namely, echo and fsize.
The other directives and a brief description of each is as follows:
#config This directive is used for formatting the output of other directives. It
does not insert any content by itself. The #config directive has several
parameters, errmsg, timefmt, and sizfmt these can respectively set alternative
error message, format in which the date and time is displayed, and the format
and unit in which file sizes are displayed. In the following example the first
line sets an alternative error message, if any subsequent directive has an error,
instead of displaying ‘An error occurred while processing this directive; the
system will give a friendlier message ‘Well, the file does not exist’. The
second directive ensures the file size will be displayed in bytes.
<!--#config errmsg=”Well, the file does not exist”-->
<!--#config sizfmt=”bytes”-->
#echo Displays the content of environmental variables discussed in the CGI
section and the additional ones listed earlier. The syntax is as follows:
<!--#echo var=”variable_name”-->
#exec This directive inserts the result of an external program in the
document. The exec directive can invoke any regular executable, including
system command as well as CGI scripts. It uses the cmd parameter to launch
normal executables and systems commands and the cgi parameter to launch a
CGI script. The following example lists all the users logged on the server
machine.
<!--#exec cmd=”/usr/bin/who”-->
Sometimes, it may be desirable to include output of a CGI script in the
page. It can be accomplished by the cgi parameter, as shown here.
<!--#exec cgi=”/cgi-bin/test.cgi”-->
#flastmod It displays the last modification date of a file. The file name is
supplied as the value of the parameter. The syntax is as follows:
<!--#flastmod file=”filename”-->
#fsize This directive displays the size specified in the file parameter, in bytes.
The syntax of the directive is as follows:
<!--#fsize file=”filename”-->
#include This directive is used for inserting the contents of a text file directly
into a document. For example, if a series of documents in a web site, were to
be given common footer information, rather than adding the same HTML
code to each one of the documents, that may have lead to maintenance and
modification problems, a common file called “footer.txt” is included at the
end of every HTML document on that site. The directive has two parameters:
file and virtual. The file parameter is used in the include directive, to specify
the location of the file being included relative to the document in which the
directive appears. The virtual parameter specifies the file name relative to the
web server’s document root directory. The syntax of the include directive is
as follows:
<!--#include file=”filename”-->
<!--#include virtual=”/docs/filename”-->
The common footer file can be added by embedding the following
directive just prior to the </body> tag in each document.
<!--#include file=”footer.txt”-->
In addition to these command directives, the APACHE web server also
supports an extended version of the SSI (XSSI). The extended directives are
available in the mod_include module. These contain XSSI directives to
define and assign values to variables and conditional statements. XSSI
provided directives enable users to create powerful server parsed HTML
documents.
The Server Side Includes (SSI) offers a shortcut to CGI for creating
simple dynamic HTML documents at a lower operating systems overhead. It
does not have the same level of interactivity and capability when compared
with CGI.
Active Server Pages
The Active Server Pages (ASP), developed and supported by Microsoft’s
Internet Information Server (Web server), is a HTML tag based architecture
that offers a framework for creating web based applications using HTML,
Scripts and ActiveX server components. In the ASP framework, HTML
documents have embedded scripts within the page. The server processes
these ASP documents, interprets the scripts, and the output of the script is
included in the document. ASP supports VBScripts, JAVAScripts and
Perlscripts. The server compiles ASP pages on the “fly, to service a request,
the resulting output is an HTML document that can be displayed on any
browser. For better efficiency, the web server usually compiles the ASP code
on the first download request and then stores it. The server uses the compiled
code for each subsequent request. The server recompiles the ASP code and
restores it any time the ASP source code changes. The caching of the
compiled ASP code results in improved performance.
In addition to the embedding of scripts in HTML pages, the key piece of
the ASP framework is the ActiveX component. There are a host of ActiveX
controls that can be downloaded and executed in the Microsoft browser.
Similarly, there are many ActiveX controls that offer functionality to the web
server. The ASP documents interface with these ActiveX components on the
server side. For example, the ActiveX component Active Data Object (ADO)
can be invoked to act as an intermediary between the Active Server Pages
and relational databases. The ADO provides many objects that are used for
connecting to databases and manipulating data. Active Server Pages can
utilize the ADO and other ActiveX components to connect, retrieve and store
data from various data sources.
As described earlier web servers interpret Active Server Pages and the
resulting pages made up of HTML are delivered to browsers. The pages
received by the browser contain normal HTML code. Fig. 7.14 depicts the
interaction amongst the various components for fulfilling requests for Active
Server Pages.
Fig. 7.14 Interactions in Processing of Active Server Pages
A browser request for the active server page, usually identified by the
“.asp” extension, is received by the web server. The web server invokes the
active server page engines if the requested file has “.asp” extension. The
server checks whether the document is being requested for the first time after
modifications, in which case, the document is parsed, syntax-checked and
compiled by the web server. If it is a repeat request, the compiled version is
loaded from the cache for improved performance. In the parsing process the
HTML code and script components are separated, the server checks for the
HTML portion and the script components are handed over to the appropriate
script engine for checking and validation. The relevant script engine executes
the script code. During the execution, script engines use the resources of the
web server. Objects that the language engine cannot handle are handed over
to the Internet Information Server (IIS), which, in turn, handles the input and
output for the ActiveX components. For unknown objects the IIS generates
an error message. At the end of execution, the script output replaces all the
script codes in the original ASP document. The resulting document,
containing only the HTML code, is delivered back to the browser for
rendering.
The Apache Mod_Perl Module
The mod_perl module offers a programming interface approach for reducing
much of the overheads associated with the CGI approach. The Apache web
server, described in earlier chapters, has been constructed using a group of
programs called modules. A request received by the Apache web server is
passed through several modules for processing, each one checks if it is
expected to handle it. Writing and adding newer modules can extend the
Apache web server functionality. The process of writing a newer module and
adding it to the Apache web server requires little more understanding than the
brief introduction provided here. The approach offers a powerful and efficient
mechanism to users, where the programs run as a part of the web server. But,
writing such programs is a not a trivial task.
The Apache web server offers a specific module called mod_perl that
embeds the Perl interpreter inside the Apache web server. When a user
request is passed to the mod-perl module, it checks the Apache registry to
determine whether it is responsible for processing the request. Through the
Apache::Registry Module, the Apache web server can be configured in such
a way that it identifies files, with certain extensions of files residing in certain
directories, for execution by the mod_perl module.
Thus, the mod_perl module enables Perl script writers to run their scripts
within the Apache web server itself, without invoking the external process
and Perl interpreter as is the case with CGI/Perl. The option of running
scripts through the embedded perl interpreter, instead of regular perl
interpreter, reduces performance overheads and also offers perl script writers
the capability to access and interface with the Apache’s Application
Programming Interface (API).
Dynamic HTML
The use of CGI and its alternatives enable the dynamic creation of web
content (pages) on the web server. These technologies provide web servers
with the capability to handle information request, act as gateway to the data
storage systems and generate the dynamic document as a response. The
response document, generated by the web server, creates dynamic content
that is delivered to the browser. Although created dynamically, the document
remains static on the browser. As a result, the content or the layout of a
document displayed on the browser cannot be changed without going back to
the server for accessing it. The standard HTML, a static language with scant
concern for the layout and style of the rendered document, does not provide
inbuilt features for dynamically updating the content, changing the
appearance and hiding, or animating the content. Cascading style sheets were
introduced to address layout related concerns. Dynamic HTML (DHTML)
provides the capability to change the HTML page even after the browser has
rendered it. For example, an image rendered on screen may change to an
alternate image on moving the mouse over it or the header of an important
text may scroll horizontally over the screen.
DHTML by itself is not a tagging language, a technology like JavaScript,
or even a plug-in. Instead, it is a concept that has been enabled by a number
of technologies such as client side scripting languages (JavaScript,
VBScript), Document Object Model, Cascading Style Sheets (CSS), and
Layers. The incorporation of these technologies enable browsers to identify
events such as passing of the cursor over an object on the document and
initiating an action, resulting in a change in the displayed document. The
concept of DHTML is achieved by marrying HTML, Cascading Style Sheets,
scripting language (JavaScript), and the Document Object Model (DOM)
together.
Client-side scripting languages provide the ability to add event driven
programming on the browser. Languages such as JavaScript can be
embedded in the HTML code with the <script> </script> tag pair. The
JavaScript provides powerful mechanisms to detect events and initiate actions
of various objects. It can be used for providing interactions with various
objects within the HTML documents or for offering some stand-alone
computations. The Microsoft Internet Explorer and Netscape Navigator,
version 4.0 and above, make a rich set of HTML elements accessible to the
client-side scripting languages. The access to these HTML elements has been
defined in the DOM.
The document object model is the heart of dynamic HTML. It is the
document object model that makes various elements of the document
accessible, thus permitting dynamic changes in the HTML. All HTML
elements such as forms, fields, images, and anchors are organized in a
hierarchical fashion, with the document object, at the top of the hierarchy.
Also, the various attributes of the browser object, windows object, document
object, various HTML element objects, and environmental information such
as date/time makes up the DOM. The browsers, by exposing the DOM to the
scripting language’s environment, offer an opportunity to manipulate these
objects and their attributes. The client side scripting languages can change the
attribute values for any DOM exposed object. This provides interactive and
dynamic web pages that can be changed by the client, even after rendering.
The Cascade Style Sheets are used for describing the layout of a HTML
document. The CSS offer a mechanism to control the rendering of the HTML
elements on a browser, without compromising the structure. They are used
for defining fonts, colors, typefaces and other styles. The style sheets act
much like templates in desktop publishing (DTP) applications. They specify a
set of conditions for rendering various HTML elements, by describing how a
document should be presented on the screen. The CSS puts the typographic
controls in the user’s hands by allowing control over the positioning of the
HTML elements and the fonts to be downloaded dynamically. The CSS are
also part of the DOM and hence all its properties are accessible to the client
side scripting language. Therefore, it is possible to change anything about the
style and the look of a page on a browser.
In short, in dynamic HTML the client-side scripting languages, through
the exposure provided by the DOM, change the elements of Cascading Style
Sheets (CSS) or the properties of the HTML elements.
HTML Editors
In the preceding section, we briefly introduced static HTML, dynamic
content creation using CGI, and alternatives to CGI, and finally discussed the
creation of interactive web pages, using the concept of dynamic HTML. The
introduction touched upon only a limited subset of HTML tags, CGI options
and alternatives. To write rich documents with style sheets and dynamic
HTML components, a greater and more comprehensive exposure may be
desirable. Writing HTML documents using a text editor the requires
knowledge of a variety of tags and attributes associated with them. With all
the matching of tag pairs and beginning and closing of quotation marks,
creating web documents for a large project becomes unwieldy.
Web authoring tools address these problems by providing an editor
environment where the HTML tags are automatically generated by the
authoring tool. The available authoring tools generally fall into three
categories;: (a) WYSIWYG editors, where you do not need to know the
HTML tags; (b) Code based editors that require basic understanding of
HTML; and (c) Compound WYSIWYG and Code based editors that can be
used by both the knowledgeable and amateur authors.
The WYSIWYG (What You See Is What You Get) editor offer an
interface that resembles the desk top publishing (DTP) graphical user
interface. The user can design a web page without knowledge of HTML,
much like word processing, by selecting and applying various options and
tools available on the interface. Editors generate HTML code in the
background. The generated HTML code tends to be complex and
cumbersome, but the user does not have to read it, unless it has to be
modified later in a different environment. The pages that have been loaded on
to the remote server need to be loaded back on the WYSIWG editor, for
carrying out modification with ease. The NetObjects Fusion, a WYSIWYG
editor enables non-HTML users to build professional looking web pages
quickly. It offers frame-based navigation bars and pop-up site maps to users.
The user designed pages are precisely and stylishly formatted with features
where graphics rotate or fade in. The Fusion (a WYSIWYG) editor locks the
user, due to its inherent nature, thus, making it extremely difficult to hand
edit the HTML code or easily rebuild the same site somewhere else.
NetObjects Fusion enables the creation of almost anything viewable in a
browser, including most current HTML features. It provides a tree structured
site diagram that lets users rearrange pages in a site. Everything on a
NetObjects Fusion is better suited for graphics rich rather than text-heavy
web sites, as even to type a headline the user has to click on the text tool and
drag the mouse to create a text frame. It supports frames and images, Java
applets; ActiveX controls; fields that display data from a built-in or external
database; and standard images, lines, and shapes. The fusion also provides a
customized style gallery so the overall look of a site can be changed with a
few mouse clicks.
Code based HTML editors, on the other hand, allow the designing of web
pages by offering GUI interface, which offers assistance by generating
appropriate HTML tags. In these editors, the user works directly with the
HTML tags and maintains control over the layout and organization of the
code. These editors do not alter the layout of other imported HTML
documents. The editor offers the graphics wizard to add the necessary code
for creating tables, frames, and other complex features. Most of these editors
also let you preview the work in a separate graphical window. HomeSite,
HotDog Professional, HTMLed Pro 32, WebberActive and WebEdit Pro are
some code based HTML editors. Allaire’s HomeSite HTML editor can create
complex web pages in a matter of minutes. The editor offers a drop down list
for selecting the attributes and values for any HTML tag that the designer
may type. It matches the closing tags and inserts them automatically. The
customizable interface of HomeSite can be accessed as a simple editing
window or a full fledged development environment. The customizable
toolbars consist of buttons for most current web technologies such as
Cascading Style Sheets, Java, ActiveX, Handheld Device Markup Language
(HDML), and Allaire’s companion ColdFusion Web development product.
Hybrid editors place themselves between the two extremes and offer the
best of both. Microsoft FrontPage, Adobe PageMill, HoTMetaL Pro,
Macromedia DreamWeaver, and QuickSite are good examples of Hybrid
editors. These editors offer WYSIWYG interface for accomplishing much of
the development, but provide the user with the capability to switch from the
word processor style window to the source code window for editing the
underlying HTML code. For example, Microsoft FrontPage offers a lucid
interface for creating HTML pages similar to documents created in word
processors. It also offers capability and dialog boxes to add VBScripts,
Jscripts, and attach the ActiveX controls. These GUI driven features assist the
user in quickly building impressive pages with dynamic HTML features. The
editor also offers access to the source code view of the HTML document,
which can be modified in the text mode.
MULTIMEDIA CONTENT
The Web integrates text and multimedia information on the same document
with relative ease. Graphics and multimedia information on a web page
makes the experience of browsing more appealing and interactive. Images
enhance the look of a web site and are essential for providing users with a
look and feel of the product, especially in the case of e-commerce
applications. Online shoppers are likely to feel more comfortable with rich
graphic and multimedia representations of the product in addition to
information. An accurate graphical view not only helps in attracting
customers but also reduces returned items, as otherwise the customer may
find that the item delivered is not the same as the one depicted on the web
site. Web designers generally do not have the knowledge and experience
required for professional graphic art. Good graphic design requires a great
deal of understanding of the image formats, and colors and color depths,
dithering, gamma correction, raster, and vector graphics. Once rich quality
images or multimedia content has been created it can easily be integrated on
to a web page, through HTML tags. There are plenty of tools in the
marketplace that can create rich multimedia content and graphic images. But,
the images may not offer a great experience to the viewer of the page due to
download time and poor rendering of images by web browsers. In order to
offer a great overall experience to the user, attention needs to be paid to
resolution, download time, format, browser compatibility, scalability, and
backward compatibility. Since images dominate multimedia content on the
web, they are mostly rendered online. The next section is devoted to
understanding the various aspects and formats of graphics content.
Graphics/Images
Higher resolution images offer a richer experience at the cost of higher
download time, due to larger file sizes. Web site designers have to strike a
balance between image resolution and download time, so that the amount of
waiting time for the user does not become excessive. There are a variety of
image formats, some of these can be rendered online by the browser but
others may require external plug-ins. The awareness of various image formats
and animation tools is an important aspect in creating images for web pages.
It is important that the image formats should be compatible and supported by
the variety of browsers prevalent in the marketplace. In other words, the
format selected should be information rich, multifunctional, and bandwidth
friendly, requiring no additional plug-ins or display software on browsers. In
addition to serving pages to a variety of browsers such as Netscape, Internet
Explorer, and Cello, the site may also have to serve pages to various versions
of a browser (e.g., IE 3.0, IE 4.0, and IE 5.0). It is important for the page
designer to employ an image format which is compatible with the older
versions as well. As images are an essential element of web pages, the basics
of digital images and images on web are described in the following
paragraphs.
The images that appear on computer monitors are a collection of pixels in
different colors. Computer monitors in essence operate with three basic
colors—Red, Green and Blue (RGB). A full range of the hues and tones of
these colors are derived by mixing various intensities of light in the three
basic colors, in each pixel. Thus, three numbers representing RGB (8-bit
colors) with values ranging from 0 (dark) to 255 (full strength) denote each
color in the RGB scheme. For example, R=255, G=0, and B=0 denotes pure
red color. Today, computers represent colors in upto 24 bits. The number of
bits that are used for representing a color is also called color depth.
Obviously, with higher color depths, its possible to represent a larger number
of colors and hues. Color depth is important from two aspects. First, the
monitor’s color depth, which is governed by the hardware and display
drivers. Typically, the operating system provides a control to configure the
color depth of the monitor within the range supported by the hardware.
Second, the color depth that is used for storing information on image files.
The color depth of the image file depends on the format in which the file is
stored. Today, the RGB uses three 8 bit channels adding up to 24 bits of color
information. This 24-bit color is also called True Color. A true color monitor
displays pixel colors exactly. The option to configure the monitor in true
color mode is often available as ‘Million Colors’. Similarly, the true color
image file records colors precisely. The human eye can distinguish only a
limited number of colors and a far lower number of hues. Thus, from the
human eye’s perspective, the picture may look as good even with lesser color
depth. Many computer systems offer a 16-bit color depth scheme that can
represent thousands of colors. It is also called the high color scheme. In this
scheme, the red color uses 32 levels (5 bits), green uses 64 levels (6 bits) and
blue uses 32 levels (5 bits), making up 16 bits of color depth. The high color
with insignificant noticeable visual differences boosts the video performance
significantly. Most of the systems use 24 bits depth for image storage but
round it off to a 16 bit color scheme at the time of displaying it on the
monitor. This ensures that the stored image retains the true colors and can be
used either way.
Raster and Vector Formats
The image files maintain information about the pixel color map that appears
on the monitor. As the images on monitors are a collection of colored pixels,
the image files can store the colored pixels quite literally. In this format the
images can be edited and modified by a bitmap editor. This format of storing
images in terms of pixels is also called the raster image format. A raster
format uses one or more bits to store a pixel information. The number of bits
used for storing single pixel information depends upon the color depth. If
only a single bit (color depth 1) is used for each pixel, it will be a black and
white image as the pixel can have only 1 or 0 value. With 8 bits for each
pixel the image can have 256 colors and as stated earlier with 24 bits for each
pixel it can have millions of colors. But obviously, the higher the color depth
(number of bits/pixel) the larger the size of the image file. The three common
internet image formats GIF, JPEG, and BMP are examples of the raster file
formats. Bitmap (BMP) files are larger in size and are used rarely on web
pages, on the other hand GIF uses only 8 bits per pixel and JPEG uses
compression technique, to reduce the size of file, and both are commonly
used in web pages.
The vector format on the other hand records images in a file descriptively,
in terms of geometric shapes. At the time of rendering on the monitor these
shapes are converted into bitmaps. Since the images in vector format are
made up of multiple independent shapes, it is easier to modify a vector
image. The component shapes of an image can be resized, rotated, moved or
even deleted independently. Postscript describes images in vector formats.
Macromedia Flash also uses the vector format for storing images on a file.
Changes in raster files are possible through modification of pixels, but this
can become cumbersome and time consuming. For example, suppose a piece
of text appearing in an image needs to be reworded, in the raster image
format, all pixels that form the text have to be modified. In vector image
formats each component can be individually selected and modified for the
new text. In this format, since the information encoded in vectors, the image
can be expanded up or down without any loss in the quality of the picture.
Vector formats provide scalable images that do not look jagged on scaling up,
or crowded on scaling down. The WWW Consortium (W3C) has developed
and is promoting a Scalable Vector Graphics (SVG) format for images. SVG
is a XML based format that can be used for describing two-dimensional
graphics. It is capable of describing vector graphic shapes, images and text
and transforming them. It can group, alter, compose, and transform these
objects together. SVG objects can be animated decoratively or by scripting.
The SVG Document Object Model (SVGDOM) offers access to all the
elements that make up the object, leading to sophisticated animation by
scripting languages.
True and Web Images
Quality images typically use 24 bits for colors (true colors). These images,
also called true images, record colors at their finest levels. True images are
useful when constructing and editing images, as they lose very little or no
information. These images, due to the 24 bit color information for each pixel,
are usually large in size. Thus, using them on web pages tends to slowdown
the download speed of the page. Moreover, monitors and human eyes may
not be able to distinguish such fine color hues. On web pages, for better
performance, it is important to have image files with smaller file sizes. GIF
and JPEG are two common formats that are used for images on web pages.
Both of these formats have smaller sized files as they compromise on the
image quality through compression. So, if the image needs to be re-edited at
some later point, it is important to keep a copy of the image in true format.
True Image Formats
As stated earlier, images in the true image format are stored accurately for
future editing. There are a variety of true image formats and each operating
system supports at least one of them as its native image format. All the
applications available on these operating systems support the native format.
Microsoft Windows uses BMP, Macintosh PICT and X-Window systems
favor XWD for true images. These formats store colors in full 24 bits but
have the capability to compress them to 16, 8, 4 and even 1-bit format. For
cross-platform applications Tagged Information File Format (TIFF) and
Portable Network Graphics (PNG) are often used.
The TIFF is a loss free, full 24-bit color format supported by many
applications for cross-platform use. The format was designed in the 80’s for
sharing and porting graphics across various platforms. It supports color depth
of 24 bits and can store photographic images well. TIFF files can be edited in
leading graphic/image editors like Adobe Photoshop and CorelDraw.
The Portable Network Graphic (PNG) is a true image format that supports
24-bit, 32-bit and even 48-bit color depths. It compresses better than GIF but
without losing image information. GIF uses 8-bit color depths and at 8-bit
color depths the PNG file tends to be 10–30% smaller in size. In PNG, 8 or
16 bits alpha channel offers varying degrees of transparency from completely
transparent to opaque images. The alpha channel lets the images appear
seamlessly over any background. The internal support for gamma correction
in PNG images provides cross-platform control of image brightness. Images
created on Macintosh offer identical appearance on a Windows platform.
Almost all the major browsers support the format. Graphics/image editors
like CorelDraw, MS Image Composer, Macromedia Freehand, and Xpaint
support editing of PNG files.
Web Image Formats
Images are an important element on the web page that enhance its looks and
attractiveness. In addition, at times, images can communicate information
more effectively than textual description. As and large effective images tend
to slow down the loading of a web page, they may render the site unusable
and deter the people from visiting again, if not included with due care. Thus,
the web page designer has to strike a delicate balance between the download
time and quality of an image. True images with 24 bits of color depth are
excessively large and create a significant delay in downloading. As a result,
the images used on web pages are generally in compressed file formats. GIF
and JPEG are two commonly used compressed formats used on the web. The
PNG format with lower color depth is also used on the web pages.
Graphic Interchange Format (GIF)
The Graphic Interchange Format (GIF) reduces the size of a true color or bit
mapped file by compressing it. It uses the Lempel-Ziv compression
algorithm. The algorithm treats rows of the same color pixels as a single unit
and saves on space. GIF uses an indexed color scheme that uses 8 bits of
color depth to index a 24-bit color palette. Thus, a GIF file can have
maximum of 256 colors in an image. This is the reason it does not work well
with photographic images or images with large number of colors. If the
image is limited to 256 colors, it performs well. Further, with a lower number
of colors it can realize even greater compression as it can reference 128
colors with 7 bits, for 64 colors with 6 bits, 32 colors with 5 bits, and 2 colors
with 1-bit. GIF is well suited for simple drawings with few colors, adding and
removing colors in a GIF file impacts the size of the file.
Conversion of full color depth images to GIF or reduction of an existing
GIF file requires reduction in the number of colors. Image editors that
support saving in the GIF format contain options for reducing colors. These
options may appear in the menu as indexed colors, reduced colors, 256
colors, or 8-bit colors. Reduction in the number of colors in an image, with
continuous tones such as photographs, may reduce the quality of the image
substantially. In such a situation editors/browsers can use the dithering
option. Dithered images create an illusion of more colors by dithering the
available hues in a diffuse pattern of pixels, in order to approximate the
original color. The other option is color substitution using the closest possible
color available, in the reduced palette, for the original color in the image.
GIF files also support transparency of images. In a transparent image the
page background shows through the background of image. For example,
assume that an image with a black background is being displayed on a page
with a white background. The image will appear on the page surrounded by
the black background frame, while in case of transparent images there will be
no such bordering frame, in stead the white background of the page will show
in place of the black background area (Fig 7.15)
SUMMARY
The exponential growth of the world wide web can be attributed to its ability
to seamlessly integrate multimedia information in a distributed environment.
Much of it is accomplished through the use of HTTP and HTML. Web
servers serve documents written in HTML to the browsers. The browsers are
responsible for interpreting these documents and rendering them at the client
site. HTML is a markup language from the family of Standard General
Markup Language (SGML). HTML is made up of text formatting, block
structuring, list, hyperlinking and other media related tags for publishing
hyperlinked multimedia documents.
The web also offers the capability to execute scripts/programs on the
server and deliver the output to the client requesting it. Functionality is
provided through a mechanism called common gateway interface (CGI). The
common gateway interface defines the input and output specification for
programs that are executed through CGI. The form mechanism of HTML is
used for presenting a form to the client, where the user can input data. The
name of the script/program (URL) is specified as an attribute of the form tag.
The script name specified by the URL is executed at the machine/server
specified by the URL, using the data entered by the user in the various fields
of the form.
The common gateway interface executes the program each time it is
requested by the client, incurring process creation overheads several times
over. Various alternatives that offer similar functionality in an integrated
fashion have also been in use. Server Side Includes (SSI), Active Server
Pages (ASP) and Apache mod_perl module are few of these alternatives.
The CGI and its alternatives provide the ability to serve dynamic content
to browsers. But, the content once delivered to the browsers, remains static to
the extent that even the style and layout cannot be changed without going
back to the server. Dynamic HTML (DHTML) extends HTML to address
these concerns. It offers the capability to change the contents of a page even
after the browser has displayed it. Dynamic HTML relies on Document
Object Model to make all the HTML elements accessible.
Content development for the web can be done using WYSIWYG HTML
editors. These editors offer graphical user interface for creating and
formatting HTML content and permit easy interface for integrating
multimedia objects on a web page. Typical multimedia content on the web
consists of images, and audio and video information. The images on the web
can be of raster or vector formats. The downloaded web content, includes
images that are usually in compressed formats such as JPEG, GIF, and PNG.
The web can also offer integrated Virtual Reality content by integrating
VRML within the environment of the world wide web.
REVIEW QUESTIONS
1. What is difference between the ‘get’ and ‘post’ methods?
2. What is difference between #include and #exec in Server Side Includes?
3. What is implication of parsing the entire HTML documents for SSI?
4. Write a HTML document that accepts the name and address of a person
from a form.
5. Write a CGI program to echo back the name and address entered by user
and attach it to the above form.
6. What are three major parts of Active Server Pages?
7. What are differences between CGI and mod_perl?
8. Write a HTML document using SSI for printing the current dates.
9. Briefly describe VRML.
10. What are raster and vector graphic formats?
11. What is a True Image format? Compare it with web image formats.
12. Write a HTML document that plays background music when loaded on
a browser.
REFERENCES AND RECOMMENDED READINGS
1. Berlin, D. CGI Programming Unleashed, Sams Publishing (1996).
2. Boutell, T. CGI Programming in C and Perl, Addison-Wesley Reading,
MA. (1996).
3. Comer, D. E. Computer Networks and Internet’s 2nded. Upper Saddle
River, NJ. Prentice-Hall (2000).
4. Deitel, H. M., P. J.Deitel and P. R.Nieto, Internet and World Wide Web
—How to Program, Upper Saddle River, NJ. Prentice-Hall( 2000).
5. Hall, M. Core Web Programming, Upper Saddle River, NJ. Prentice-
Hall (1998).
6. Hobuss, J. J., Building Access Web Sites, Prentice-Hall Upper Saddle
River, NJ; (1998).
7. Jardin, C. A. Java Electronic Commerce Sourcebook, New York: John
Wiley and Sons (1997).
8. LeMay, L. Teach Yourself Web Publishing with HTML4, Sams
Publishing (1999).
9. Walther, S., J. Levine, Teach yourself E-commerce Programming with
ASP, Sams Publishing (2000).
Learning Objectives
This chapter covers the following topics:
1. Importance of security for Electronic Commerce and the inherent
vulnerability of the Internet
2. Security Policy, Procedure and Practices
3. Site Security
(a) Sources of vulnerability, types of attacks and prevention
(b) Fortifying the access through firewalls
(c) Various firewall configurations
4. Protecting the Web (HTTP) Service
(a) Server privileges
(b) Protecting confidential resources on the site
(c) Vulnerability of Common Gateway scripts and preventive measures
The internet offers tremendous cost savings and productivity gains, as well as
significant opportunities for generating revenue, to the business. However,
along with the convenience and easy access to information come new risks.
Among them is the risk that valuable data or information may be lost, stolen,
corrupted, or misused. Information recorded electronically, and available on
networked computers, is more vulnerable compared to the same information
being printed on paper and locked in a file cabinet.
In the increasingly competitive environment, an unscrupulous competitor
may try deriving advantage by intruding and getting access to his
competitor’s financial, design and other transactional information. Cyber
intrusions between Indian and Pakistani hackers, assaulting and defacing web
sites controlled by the other side, and Taiwanese and Chinese hacking into
sites supporting viewpoints, other than their are some common examples of
this vulnerability. The web sites of Bhabha Atomic Research Center (BARC),
National Informatics Center (NIC) of India, Microsoft, NASA, Whitehouse,
FBI, CNN, eBay, and Amazon have all have been hacked and defaced by
intruders at one point of time or the other.
A cyber intruder does not need to break into an office or home, and may
not even be in the same country. The intruder can steal or tamper information
sitting in the comfort of his own room. The intruder can create new programs
and run them on remote computers, causing the system to malfunction, while
hiding evidence of his unauthorized activity. Additionally, in the
transactional world of electronic commerce, the information transmitted over
the network can be tapped and tampered with.
In the internet based business environment, business and transaction
information is hosted on a site that runs services such as web and mail. Thus,
comprehensive handling of the security of an internet based business requires
addressing the security issue at the following three levels:
1. Site Security—Security of the host computer
2. Services Security—Security of information distribution services such as
HTTP servers, SMTP servers, FTP servers
3. Transaction Security—Since the transaction information travels over the
wire, it needs to secured from intruders trying to access and comprehend
or tamper with it.
As safeguarding the environment requires resources, the higher the degree
of security requirement the larger the resource cost is likely to be. Thus, it is
important to assess the level of protection an organization can afford, or may
truly require. The information is collected by carrying out a risk analysis of
assets that require protection, like the network, volume traffic ahead
information and transaction, as well as factors like the likely attackers,
immediate cost of compromise and recovery from the attack. The information
is used for assessing the level of protection and the areas of vulnerability and
thereby developing the security policy of the organization.
This chapter addresses the first two issues of securing the site and the
services from intrusion and compromise. The third issue of securing the on-
the-wire transaction is addressed in the following chapter.
SITE SECURITY
A site is any organization that has network-related resources like host
computers that users use routers, terminal servers, PCs, or other devices that
are connected to internet. A site may be service provider such as a mid-level
network or an end user of internet services. It is important that the services
hosted by the site provide the intended functionality to legitimate clients,
without any breakdown. Occasionally, a hacker may try to break-in and
disrupt the services or alter the contents of the site, which may be
embarrassing to the organization.
The following section lists the issues and factors involved in securing the
services and the network at the site location.
Separation of Services
A site may wish to provide many services to its users, some of which may be
external. The services may have different levels of access needs and models
of trust. Apart from performance reasons, there are a variety of security
reasons to attempt to isolate the services onto dedicated host computers.
Services which are essential to the security or smooth operation of a site
would be better off being placed on a dedicated machine with very limited
access, rather than on a machine that is used for providing greater
accessibility and other services that may be prone to security lapses.
There are two conflicting, underlying philosophies that can be adopted
when defining a security plan. The choice between them depends on the site
and its needs for security.
1. The “deny all” model suggests turning off all services and then
selectively enabling services on a case by case basis as required. This
can be done at the host or network level, as appropriate. This model is
generally more secure than the next one. However, more work and a
better understanding of services is required to successfully implement a
“deny all” configuration.
2. The “allow all” model is based on the logic of simply turning on all
services, usually with the default at the host level; and allowing all
protocols to travel across network boundaries, usually with the default at
the router level. As security gaps become apparent, they are restricted or
patched at either the host or network level. This model is much easier to
implement, but is generally less secure than the “deny all” model.
Each of these models can be applied to different portions of the site,
depending on factors like functionality requirements, administrative control,
and site policy. For example, an “allow all” policy may be adopted for traffic
between a LAN’s internal to the site, but a “deny all” policy can be adopted
between the site and the internet.
FIREWALLS
A firewall is a controlled access point between security domains, usually with
different levels of trust. It acts as a gateway through which all traffic to and
from the protected network and/or systems passes. It helps to build a wall
between one part of a network and another part. For example, placing
limitations on the amount and type of communication that takes place can
separate a company’s internal network and the internet. The unique feature
about this wall is that there needs to be way for some traffic, with particular
characteristics, to pass through carefully monitored doors (“gateways”). The
difficult part is establishing the criteria by which the packets are allowed or
denied access through the doors.
Firewalls can be a highly effective tool in implementing a network
security policy if they are configured and maintained correctly. They provide
a certain level of protection and are, in general, a way of implementing
security policy at the network level. The level of security that a firewall
provides can vary depending on the level of security required on a particular
machine. There are other considerations as well, like the traditional trade-off
between security, ease of use, cost, and complexity.
Types of Firewall
Firewalls can have variety of configurations, depending upon the security
requirements and availability of resources for a site. Broadly speaking, there
are four types firewalls which accomplish controlled access, using following
methods:
1. Packet Filtering
2. Circuit Level Gateway
3. Application Level Gateway
4. Stateful Inspection
Firewalls that are commercially or publicly available, employ a
combination of these four key capabilities to ensure a secure environment.
The key capabilities can be used for differentiating and evaluating the
effectiveness of a given electronic commerce environment.
Packet-Filtering Firewall
A packet-filtering firewall operates by filtering the incoming and outgoing
packets, using the router or devices that have been configured to screen
incoming and outgoing packets. It examines the information contained in
TCP and IP packet headers, in order to accept or deny packets from entering
or leaving the network. The examining filters can be configured to accept or
discard a packet, based on the packet’s full association, consisting of the
following attributes:
Source address
Destination address
Application or protocol
Source port number
Destination port number
All the routers examine packet headers, to determine the source and
destination address contained in the packet. In consultation with the routing
table, the routers determine the next hop of the arriving packet. The packet is
forwarded to the line that leads to the next hop of the packet. A packet-
filtering firewall is a router that goes a step further. These routers store a table
containing rules specified for security purposes. The router, during
examination of the attributes contained in the packet header, compares them
with the rules stored in the “access control” table. The rules dictate whether
the firewall should discard the packet or permit the packet to pass through the
router.
A packet-filtering firewall reads the packet header and scans the rules
table for a match, if it finds a rule that matches with the information
contained in packet, it takes the action specified in the rule. If the information
contained in packet does not match against any of the specified rules, the
firewall applies the default rule. It is necessary to specify a default rule
explicitly in the firewall’s table. The default rule generally follows the “allow
all” or “deny all” model. For strict security, the firewall default rule should
follow the “deny all” model, which instructs the firewall to drop a packet that
meets none of other the specified rules in the table.
In the packet-filtering firewall, one can define packet-filtering rules that
specify which packets should be accepted and which packets should be
discarded. For example, the rules configured could be to permit all the traffic
to pass through except from some “untrusted” servers specified by their IP
addresses. Or, the “deny all” model can be adopted as the default rule,
permitting packets only from a list of trusted servers, specified by their IP
addresses. Filtering rules can specify the packets other than those with
destination address of the mail server will not be permitted. In addition, even
the mail packets meant for the mail server, from hosts that may have mail-
bombed the receiving server in the past, will be discarded.
The packet-filtering firewall can be configured to screen, not only IP
packets but, packets based on TCP and User Datagram Protocol (UDP) port
numbers as well. Rules that screen the port number can be used for
configuring a firewall, that enables the specification of the different types of
connections that can be accepted. A firewall can be configured to accept only
mail and WWW connections coming from outside hosts, by specifying a rule
that permits the traffic meant for mail server i.e., SMTP (port 25) and web
server i.e., WWW (port 80). However, these rules will be able to filter the
traffic, provided the servers follow a TCP/IP network convention—servers
(and clients) generally run particular TCP/IP applications over particular
ports (often referred to as well-known ports).
Packet-filtering firewalls provide a measure of protection at a relatively
low cost, and with very little or no delay in network performance. Creating a
packet-filtering firewall requires an IP router with packet-filtering capabilities
to which packet-filtering rules can be added at no extra cost. Today, most IP
routers manufactured by Novell, Cisco Systems, and Bay Networks are
capable of filtering incoming and outgoing packets.
The creation of packet-filtering rules can become tedious when used for
filtering all the permutations and combinations of packet attributes.
Assuming that the router has been equipped with effective rules, a packet-
filtering firewall still has inherent limitations and cannot deter hackers with
more than a passing interest in your network. For example, if a rule instructed
the firewall to drop incoming packets with unknown source addresses, it will
block hackers from accessing trusted servers on the network. But, a seasoned
hacker can substitute the actual source address on a malicious packet, with
the source address of a trusted client and yet gain access.
In short, packet-filters have the following advantages:
Packet filters tend to be very fast and tend to be transparent to users.
Packet filters can be very effective in completely blocking specific types
of traffic, and for this reason are sometimes part of an overall firewall
system. For example, applying a filter to discard packets for TCP port
23 (Telnet) can easily block Telnet.
However, packet-filtering firewalls also have the following limitations:
For useful and effective filtering, filtering rules lists can become
lengthy, quite complex and error-prone. Although, performance is not
usually a severe problem in new router implementations, lengthy access
lists can degrade throughput and increase latency. In a packet filtering
router, every packet going through must be checked against the same
access lists as it does not maintain state information.
Packet-filter cannot support user authentication and blocking based on
contents at the application level.
For complex protocols that specify return data ports dynamically, the
filtering protocol becomes difficult and complex.
Circuit Level Firewall
A circuit level firewall operates at the session layer level of the OSI model. It
relies on TCP session layer protocol and monitors TCP handshaking between
packets, from trusted clients or servers to untrusted hosts and vice versa, to
determine whether a requested session is legitimate. In other words, it means
that the firewall doesn’t simply allow or disallow packets, but also
determines whether the connection between both ends is valid, according to
configurable rules. On validation, it opens a session and permits traffic only
from the allowed source and possibly only for a limited period of time. The
validity of the connection can be based on the following attributes:
destination IP address and/or port
source IP address and/or port
time of day
protocol
user
password
It validates each session of established connection for the exchange of data
between two machines. Circuit level filtering takes control a step further than
a packet-filter. One of the major shortcomings of a packet-filtering firewall is
that the source address is never validated, thus, an attacker can forge packets
with the permitted source IP addresses.
Circuit level firewalls determine the legitimacy of a session by checking
the connection-requests attributes against the configured filtering rules,
followed by closely monitoring the TCP handshaking process that follows the
request for opening a connection from an untrusted host. The handshaking
involves an exchange of TCP packets that are flagged SYN (synchronize) or
ACK (acknowledge). These packet types are legitimate only at certain points
during the session. A circuit-level firewall determines that a requested session
between trusted and untrusted machines is legitimate only if the SYN flags,
ACK flags, and sequence numbers involved in the TCP handshaking are in
logical sequence.
Once a circuit level firewall ascertains that a requested session is
legitimate, the connection is established. It maintains an entry for each
established connection that is active. From this point onward, the firewall
simply copies and forwards packets back and forth, with no further filtering.
The copy and forward services are performed by specialized applications,
that establish a virtual circuit or Unix-like pipe between two networks. Once
the session is closed, the firewall removes the associated entry from the
connection table, deallocating the circuit used for copying and forwarding
packets, for this connection.
A circuit level firewall also provides the capability of proxying IP
addresses. In this configuration, the circuit level firewall uses a process called
address translation, to map all the internal IP addresses to one “safe” IP
address. This address, associated with the firewall, is used as the source
address by all outgoing packets originating at the internal network. Since all
outgoing packets appear to have originated from that firewall, it shields the
trusted (internal) network from direct contact with the untrusted network. The
circuit level firewall’s IP address is the only active IP address that the
untrusted network learns about, making the trusted network safer from
spoofing attacks.
A circuit level firewall has an inherent weakness. Once the legitimacy of a
connection is established by the circuit level firewall, any application can be
run over the connection as the circuit level firewall simply copies and
forwards the packets back and forth with out examining the content. An
attacker on an untrusted network could use an established connection to
possibly slip malicious packets past the firewall. The attacker could then deal
directly with an internal server, that may not be as carefully monitored or
configured as the firewall itself. To filter the application level content of
individual packets generated by particular services, an application level
firewall is required.
Application Level Firewall
The application level firewall act as a proxy for applications. It performs all
data exchanges with the remote system on behalf of the applications running
behind the firewall. As a result, it renders the computer, behind the firewall,
all but invisible to the remote system. The application firewall can be
configured to allow or disallow traffic according to very specific rules. For
example, it may permit some commands to a server but not others, it may
limit file access to certain file types, or even offer varying levels of access
depending upon the authentication level of users. This type of firewall,
typically, performs logging of traffic and monitoring of events on the host
system. It also permits setting of alarms, system alerts, or notification to an
operator, under pre- defined conditions. These firewalls are regarded highly
secure. They certainly have the most sophisticated capabilities. An
application firewall is normally implemented on a separate computer on the
network, whose primary function is to provide proxy service to various
applications.
An application-level firewall intercepts incoming and outgoing packets,
runs proxies that copy and forward information across the firewall, and
functions as a proxy server. As a result it prevents any direct connection
between a trusted server or client and an untrusted host. However, the
proxies, that an application-level firewall runs, are application level proxies
and can filter packets at the application layer level of the OSI model.
Fig. 8.1 Application Level Proxies for Services
Application-level proxies are designed for individual applications. Thus,
application-specific proxies accept packets only from the services they are
designed to copy, forward, and filter. For example, FTP proxy can copy,
forward, and filter FTP traffic/packets only. It implies that on a network that
relies on an application level firewall, incoming and outgoing packets can
access only those services for which it has a proxy running. For example, if
an application level firewall ran WWW and SMTP (e-mail) proxies, only the
WWW and SMTP (e-mail) traffic will pass through the firewall, while all
other services such a Telnet and FTP would be blocked.
An application level firewall runs proxies that examine and filter
individual application packets, rather than simply copying them and blindly
forwarding them across the firewall. Thus, it can be configured to add rules
that can filter packets, based on the content. These proxies can copy forward
and filter particular kinds of commands or information in the application
protocols. For example, the FTP application proxy can be configured to block
users from executing the put command. Thus, no user can write any
information on the FTP server.
Application level firewalls are one of the most secure firewalls available.
Ideally, a firewall should be transparent along with being secure. In other
words, users on the trusted network should not feel any difference whether
they are accessing internet services through a firewall, or without it. Most
users often experience some delays, and in some configurations may have to
perform multiple logins, before they are connected to the internet or intranet,
through an application level firewall.
In short, the significant security benefits that the application layer proxy
server offers are as follows:
It is possible to add access control lists to protocols, requiring users or
systems to provide some level of authentication before access is granted.
Smarter proxy servers, also called Application Layer Gateways, can be
written to understand specific protocols, and configured to block only
subsections of the protocol. For example, an application layer gateway
for FTP can tell the difference between the “put” command and the
“get” command; an organization may wish to allow users to “get” files
from the internet, but not be able to “put” internal files on a remote
server. By contrast, a filtering router could either block all FTP access,
or none.
Proxy servers can also be configured to encrypt data streams based on a
variety of parameters. An organization might use this feature to allow
encrypted connections between two locations, whose sole access points
are on the internet.
Although proxy servers offer better and controlled filtering than simple
packet filters and circuit-level filters, but there are several
disadvantages.
In order to implement the proxy based access protocol, clients on the
protected network must be specially modified. This complicates the
configuration and adds considerably to the network administration.
Also, since the proxies are application specific, only applications that
have proxies work.
Proxy servers operate within the environment of general-purpose
operating systems, thus, becoming vulnerable to the security loopholes
of the operating system.
The performance (throughput) of the system degrades, as the number of
connections through the proxy servers go up, because of the significant
processing overheads incurred in running and handling proxy programs.
Proxy severs introduce a lot of latency, since two separate connections
must be established before any data can be transferred. New connections
suffer a from a high connection setup time due to the “process” nature of
a proxy. Each connection requires a separate process.
Stateful Inspection Firewall
The stateful inspection firewall operates at the network layer, session layer
and application layer of the OSI model, by combining the functionality of the
packet-filtering, a circuit level, and application level firewalls. At the network
layer level, the stateful inspection firewall filters all incoming and outgoing
packets based on source and destination IP addresses and port numbers. At
the session layer level, the stateful inspection firewall determines whether the
packets in a session are legitimate, by verifying that SYN and ACK flags and
sequence numbers are logical. Finally, a stateful inspection firewall mimics
an application level firewall by evaluating the contents of each packet up
through to the application layer, and ensuring that the contents match the
rules defined by the network security policy.
Like an application level firewall, a stateful inspection firewall can be
configured to drop packets that contain specific commands. For example, you
could configure a stateful inspection firewall to drop FTP packets containing
a Put or Get command. But, the main difference lies in the fact that the
application level firewalls accomplish this by establishing two connections:
one connection between the trusted client and the firewall and another
connection between the firewall and the untrusted host. The application level
proxies examine the content and relay the information between the two
connections. It ensures a high degree of security, but introduces performance
overheads. In contrast, a stateful inspection firewall permits a direct
connection between a trusted client and an untrusted host. The stateful
inspection firewall ensures security by intercepting and examining each
packet up through the application layer of the OSI model. A stateful
inspection firewall relies on algorithms that compare packets against the
known bit-patterns of authorized packets, to recognize and process
application layer data, providing them the ability to filter packets more
efficiently than application specific proxies.
The biggest advantage in using stateful inspection firewall for securing
internet and intranet connections is the transparency it offers to users. It does
not require running proxy services, or modifying clients and hosts to go
through the proxy services for data scrutiny, even at the application level.
Most firewalls provide logging which can be tuned to make security
administration of the network more convenient. Logging may be centralized
and the system may be configured to send out alerts for abnormal conditions.
The logs should be regularly monitored, to detect any signs of intrusions or
break-in attempts. Since some intruders will attempt to cover their tracks by
editing logs, it is desirable to protect these logs. This can be achieved by
using any of the available methods: write once, read many (WORM) drives;
papers logs; and centralized logging via the “syslog” utility. Another
technique is to use a “fake” serial printer, but have the serial port connected
to an isolated machine that keeps the logs.
What a Firewall Cannot Do?
There is a general misconception that a firewall is a panacea to all security
problems. A properly configured firewall system helps in eliminating many
threats pertaining to the security of a server/site, but there are certain things,
which it cannot perform.
Firewalls cannot protect against attacks that do not go through it. In
other words if one of the servers in the trusted network supports a dial-in
access to remote users and the traffic between these machines does not
go through the firewall, it cannot offer protection against any attacks
emanating from such connections.
Firewalls do not protect against threats emanating from internal users
i.e., those who are part of the trusted network.
Firewalls are concerned with monitoring the traffic and permitting only
authenticated and legitimate traffic flow. It does not concern itself with
integrity issues related to applications and data.
For the most part firewalls, as discussed above, are concerned with the
controlled flow of data traffic and do not provide confidentiality of data.
However, application proxies at the firewall machine can provide
encryption and decryption of all the data passing through, as it becomes
a single access point to the application.
Firewalls cannot protect very well against viruses. There are too many
ways of encoding binary files for transfer over networks, as well as too
many variety of architecture and viruses making it difficult to search for
them all. In general, a firewall cannot protect against a data-driven
attack — attacks in which something is mailed or copied to an internal
host, where it is then executed.
Locating Web Server
The HTTP server, also commonly referred to as the web server, is a key
element of the electronic commerce environment. The security of the web
server is of paramount concern as all electronic commerce related
information, databases, transaction and even payments may reside or be
accessed in trusted mode from the web server.
Placing the Web Server Inside the Firewall
The most straightforward use of a firewall is to create an “internal site”, one
that is accessible only to computers within the local area network. In this
case, all that is required is to place the server inside the firewall.
The advantage of putting a web server behind a firewall is that
maintenance is much more convenient. In the internal trusted network, local
access and file updates can be enabled so that administrators and content
providers can easily log into the web server and update the content. The
firewall can be configured to block all the incoming traffic other than HTTP
requests and DNS queries. Thus, any attempts by people outside the firewall
to log into the web server and exploit security holes will be blocked because
the firewall blocks all access to the server except to the HTTP daemon and
DNS queries.
SUMMARY
Commerce over the network requires an assured level of confidence, with
regards to the security of information. Security incidences such as probes,
scans, account compromises, exploitation of trust, sniffing, and spoofing, that
are used for violating the security policy of a site, are also described. These
incidences exploit the technical vulnerability of the internet. Each business
needs to clearly spell out its security policy, procedures and practices for
implementing the desired level of security, and enabling the framework of
defense mechanism and service configurations. A comprehensive security for
electronic commerce covers security at the host level, site level, and the on-
the-wire transaction level. Site security includes detection and deterence
against sniffing and spoofing attempts and protection of important services
such as web servers, DNS servers, and other infrastructure services.
Firewalls have emerged as an important mechanism for fortifying site
security, by controlling access, monitoring, and filtering the incoming and
outgoing message traffic, right down to the packet level. Based on the
capabilities and configurations, we can classify firewalls in multiple
categories. Understanding firewalls, their capabilities, and configuration
options become important in trying to implement a security policy for a site.
The prevention of sniffing, spoofing, and access monitoring and control
through firewalls can secure a site from unwanted traffic and intrusion
attempts. In any electronic commerce environment, the commerce/web server
will receive, process and service requests from variety of unknown clients.
Thus, the server itself often becomes a prime focal point for attacks by
hackers. The vulnerability of web servers and issues that deal with the
reinforcement of security management around the web server attain prime
importance. This chapter discussed some of these vulnerabilities and how to
avoid the security pitfalls around web servers.
REVIEW QUESTIONS
1. What are network security incidences?
2. Describe what a denial of service attack is and how it affects electronic
commerce.
3. Why is the internet vulnerable to hackers? Describe various sources of
vulnerabilities.
4. What is meant by security policy? Distinguish it from security
procedures.
5. What are major threats posed by a sniffing attack?
6. Describe the important means of deterring sniffing attacks.
7. What is meant by ARP spoofing and how is it carried out?
8. What are threats posed by a DNS spoofing attack?
9. What is a firewall and how does it protects a site?
10. Briefly describe the various types of firewalls.
11. Describe pros and cons of the various ways of locating web servers in a
firewall configuration.
12. What is meant by DMZ?
13. Compare the stateful inspection firewall with the application-level proxy
firewall.
14. What are vulnerabilities of a web server?
15. Describe the important factors in planning a firewall design.
16. Which vulnerabilities of the Common Gateway Interface (CGI) can
attackers exploit ?
17. Compare compiled versus interpreted CGI scripts from the security
perspective.
TRANSACTION SECURITY
In the electronic commerce environment the transaction take place over the
network. During various phases of an electronic transaction the information
such as product specification, order details, payment and delivery information
travels over the Internet. The transaction information transmitted over the
public Internet can be tapped, intercepted, diverted, modified, and fabricated
by an intruder trying to gain some benefit or cause damages to competing
business. The intruder may be interested in seeking the confidential
information about the competing business entities or may even be interested
in misguiding to cause losses to competing business or gain benefit from such
an act. The intruding activities can be broadly classified in two categories
—passive and active intrusion.
In passive intrusion, transmissions on the network are eavesdropped on or
monitored. The motive of the attacker is to obtain the information being
transmitted. Passive attackers intercept the information, resulting in the loss
of confidentiality and privacy of the data. Passive attacks are difficult to
detect, as the data is not altered. Hence the emphasis is on prevention of such
attacks rather than detecting them. For example, data can be scrambled using
an encryption technique so that even if the intruder is able to intercept the
message, no meaningful information can be extracted from it.
Active attacks involve mutation of data or generation of counterfeit
messages. The motive of the attacker is prevent messages from reaching their
intended destination; to masquerade as another entity and get access to
restricted information; or to feed another user with falsified information, with
the aim of misleading the person. Active attacks are easier to detect as
compared to their passive counterparts. For example, a cryptographic
checksum can accompany each message. If the message is altered during the
passage in any manner, the tampering can be detected because of the
violation of the checksum. In the context of the communication over a
network, the following attacks can be identified:
Network Transaction Security Issues
Disclosure: Release of message contents to any person not authorized to
see them or not possessing the appropriate cryptographic key.
Traffic Analysis: It refers to the discovery of the pattern of traffic between
parties. In a connection-oriented application, the frequency and duration of
connections could be determined. In either a connection-oriented or
connectionless environment, the number and length of messages between
parties could be determined.
Masquerade: It refers to insertion of messages into the network, from a
fraudulent source. This includes the creating of messages by an opponent,
that are purported to come from an authorized entity. Also included are
fraudulent acknowledgments of message receipt or non-receipt by someone
other than the message recipient.
Content Modification: Changes to the contents of a message, including
insertion, deletion, transposition, or modification.
Sequence Modification: It refers to modification of the sequence of
messages between parties, including insertion, deletion, and reordering of
some sequenced packets, by the intruder, during transmission.
Timing Modification: It refers to delayed messages, or also replay of old
message sequences, that were recorded by intruder in an earlier transaction.
In a connection-oriented application, an entire session or sequence of
messages corresponding to a full session could be recorded by an intruder,
and later replayed. The destination may think of it as a valid session and
carry out the indicated transactions one more time. Also, both in connection
and connectionless services the individual messages in a sequence could be
delayed.
Repudiation: It refers to the denial of the receipt of message by the
destination or the denial of transmission of message by the source.
Security Services
In the transactional internet environment, it is important to ensure the security
of transactions as they travel over the network. As stated above, transactions
may be subjected to passive or active intrusion. Passive intrusion threatens
the loss of privacy and confidentiality of data, but and active intrusion may
result in the intruder assuming someone else identity and creating
transactions on their behalf, through fabrication. The active intruder may also
modify the content of the transaction. For example, an order being placed for
1000 items may be modified to 10,000 items that may later result in conflict
between business parties, and subsequent loss of money as well as trust. For
developing trust in the electronic commerce environment, for transactions to
take place, the following five issues are important.
Authentication
Simply stated, authentication is the process of verifying the identity of a
person from whom the communication message emanated. In the case of a
single message, authentication assures the recipient that the communication
partner is not an imposter, and that the text of the message itself has not been
altered.
In the case of an ongoing interaction, such as the connection of a remote
terminal to a host, there are two aspects of this service:
1. At the time of initiation of a connection, the verification of the two
participating entities, i.e., establishing that each of them is the same
entity what they claim to be.
2. The connection is not interfered with, in such a way that a third party
can masquerade as one of the two legitimate parties, for purposes of
unauthorized transmission or reception.
Integrity
Integrity means that it should be possible for the receiver of a message to
verify that the message has not been tampered with, while in transit. An
intruder should not be able to substitute a false message for a legitimate one.
In other words, no one should be able to add, delete or modify any part of the
message during transmission. The receiver should be in a position to verify,
in case any tampering has taken place in the message stream. The integrity of
the message prevents any intentional or unintentional modification of the
message through the use of error detection codes, checksums and sequence
numbering, time-stamping and encryption, and hashing techniques. Error
detection codes and checksums computed on fields, or entire messages, help
in detecting, and sometimes even correcting, errors that may have crept in
during transmission. Sequence numbering and time-stamping protects against
reordering, replaying, and loss of part of the message. Encryption techniques
can be used for detecting the tampering of messages. Algorithms such as
Message Digest 5 (MD5) and Secure Hash Algorithm (SHA) compute a hash
code of a fixed size, for any given message. The code computed by these
algorithms is guaranteed to be unique. In order to ensure integrity the sender
may send the message and the computed hash code as well. The receiving
side, on receiving the message, can also compute the hash code of the
received message. In case of a tampered message, the two hash codes the one
computed at receiver’s end and the one provided by sender, will not match.
Non-repudiation
Non-repudiation prevents either the sender or the receiver from denying a
transmitted message and files or data, when in fact they did. When a message
is sent, the receiver can prove that the message was in fact sent by the alleged
sender. Similarly, when a message is received, the sender can prove that the
message was in fact received by the alleged receiver. In a business
transaction, the legal framework ensures that no party is in position to
unilaterally repudiate the transaction. But, for legal purposes an agreement
should be signed by the parties. However, in the electronic commerce
environment, as transactions take place over the network, only digital
content, rather than physically signed documents, may exist.
In such a situation, let us say a customer places an order for 1000 shares of
XYZ Corporation, at Rs. 100 per share. The stock broker executes the order,
but later on the same day price drops down to Rs. 10 per share. If the
transaction was placed electronically, the customer may deny placing the
order. A similar repudiation can take place from a greedy broker, who may
discover the price for the shares have gone up to Rs. 500 per share. In either
of situation, authentication and integrity play a role, but in addition the
electronic commerce environment has to guard against repudiation by
introducing fool-proof, digitally signed contacts and agreements that can be
validated by the legal infrastructure, to offer a repudiation-free business
environment.
Confidentiality
Confidentiality is the protection of transmitted data, from passive attacks.
When a message is transmitted over the communication channel, it can be
intercepted at any point in between, through wiretapping or with the help of
computer programs. Confidentiality ensures that the contents of a message
are not leaked or revealed to a hacker as it travels to its destination. In the
electronic commerce environment, the confidentiality of payment information
and ordering details are of utmost concern. Similarly, in case of business
partners and associates sharing sensitive information over the network, a
competitor may like to have access to the information. Since, the internet
environment is quite susceptible to passive intrusion, as the packets pass
through variety of host computers, confidentiality is usually ensured by
encrypting information.
Authorization
Systems connected on the internet share information over the network,
among a variety of users. The authentication process ensures the correct
identification of the user and letting him/her in, but all the information on a
system may not be shared with all users. Authorization pertains to the
permission granted to a person or a process to do certain things. Privileges
are associated with sensitive information stored on hosts. Authentication
ascertains that the user is who he claim to be, while authorization ascertains
the rights of the claimant to access the information, before presenting the data
to him.
The confidentiality of messages in electronic commerce can be handled by
encrypting the message prior to transmitting it over the network, and finally
decrypting it at the destination. Cryptography, the science of encryption, can
be used for addressing a variety of issues related to secure communication
over the network.
CRYPTOLOGY
Introduction to Cryptography
Cryptography, or the encrypting and decrypting of messages, for sharing
secret messages among a group of users or any two persons, has existed for
thousands of years. One of the earliest uses of cryptography was by Julius
Caesar, who did not want messages carried by his couriers to fall into the
wrong hands. Caesar used a simple substitution cipher, now known as the
Caesar Cipher, to do this. Its operation was simple—each letter was rotated
by three. Thus, A became D, B became E, and so on. A generalization of
Caesar Cipher can be done by changing the rotation by 3, used in original
encryption, to k. The two persons using the systems have to know the value
of k. these rotation based algorithms are not too difficult to solve. Later better
algorithms were devised and put to use. The security of the early algorithms
depended on keeping its operation a secret, and in ensuring its restricted
usage. To ensure this, not only were the keys kept secret, but so were entire
algorithms; in order to prevent the enemy from even knowing where to start.
In modern encryption techniques, the secrecy of algorithms is a self-defeating
proposition. Instead, it is better to publicize the algorithms far and wide. So
that, any loopholes can be found. It is the key that has to be kept secret.
Cryptographic systems can be classified along three independent dimensions:
1. The methodology employed in transforming the plaintext to ciphertext.
Encryption algorithms are based on two general principles:
• Substitution: Individual elements in the plaintext are mapped into
another element or a group of elements by employing a chart or a fixed
pattern in order to disguise them. The order of the plaintext symbols is
preserved.
• Transposition: The individual elements of the plaintext are rearranged
but not disguised.
2. The number of keys employed by the algorithm.
• Symmetric, Shared-key or Conventional encryption: The same key
is shared by both the sender and the receiver, i.e., the same key is used
for encryption and decryption.
• Asymmetric, two key or public key encryption: The sender uses one
key for encryption and the receiver uses another complementary key
for decryption.
3. The manner in which the original plaintext is processed.
• Stream cipher: The individual elements of the stream of data are
processed continuously, and the output is generated accordingly.
• Block cipher: The input being processed is a block of elements, and
the output generated is a block corresponding to each input block.
Cryptanalysis
As described above, a cryptosystem or cipher system is a method of
disguising messages so that only certain people can see through the disguise.
It is usually a whole collection of algorithms. Cryptanalysis is the art of
breaking cryptosystems—seeing through the disguise even when one is not
supposed to be able to. Simply put, cryptanalysis is the process of attempting
to discover the plaintext message P or the key K or both. The strategy
employed by the cryptanalyst depends on the nature of the encryption scheme
and the information available to him. The cryptanalyst employs a variety of
methods to break the code. Typically, a cryptanalyst classifies the problem
depending upon the availability of the ciphertext or plaintext. The following
table summarizes the various scenarios that are available to a cryptanalyst.
Table 9.1 Types of Attacks on Encrypted Messages
Type of Attack Known to Cryptanalyst
Ciphertext only Encrption algorithm and ciphertext to be decoded
Known plaintext Encrption algorithm, ciphertext to be decoded; one or
more plaintext-ciphertext pairs formed with a secret
key
Chosen plaintext Encrption algorithm, ciphertext to be decoded;
plaintext message chosen by cryptanalyst, together
with its corresponding ciphertext
Chosen ciphertext Encrption algorithm, ciphertext to be decoded;
purported ciphertext chosen by cryptanalyst, together
with its correspondong decrypted plaintext generated
with a secret key
Chosen text Encrption algorithm, ciphertext to be decoded;
plaintext message chosen by cryptanalyst, together
with its corresponding ciphertext generated with the
secret key; Purported ciphertext chosen by
crptanalyst, together with its corresponding
decrypted plaintext generated with the secret key
CRYPTOGRAPHIC ALGORITHMS
DES
The Data Encryption Standard, developed by IBM, is one of the most widely
used encryption schemes. It was adopted in 1977 by the National Institute of
Standards and Technology (formerly National Bureau of Standards). It is a
block cipher based encryption technique based on a 56-bit key. DES
algorithm transforms a 64-bit input, in a series of steps into a 64-bit output.
The same steps, with the same key, are used to reverse the encryption.
AUTHENTICATION PROTOCOLS
The cryptographic algorithms discussed in previous section are used for
addressing important issues like authentication, confidentiality, integrity and
non-repudiation, that are essential for the development of electronic
commerce.
Authentication Using a Shared Secret Key
The shared secret key based authentication assumes that through some offline
or online mechanism the two parties have established a secret key. The
challenge-response based protocol, that can authenticate both parties over the
network, is based on a simple principle: one party sends a random number,
known as a challenge, to the other, who then transforms it using the shared
secret key and returns the result. The first party compares the actual response
with the expected response and verifies the identity of the second party.
DIGITAL SIGNATURES
The digital signature is to the electronic world what the handwritten signature
is to the tredeteanal/commerce.It must incorporate the following properties:
It must be able to verify the author, the date, and the time of the
signature.
It must be able to authenticate the contents, at the time of the signature.
It must be verifiable by third parties, in case of any dispute.
The above properties place the following requirements on the digital
signature:
The signature must be a bit pattern that is dependent on the message
being signed.
To prevent forgery and denial, the signature must use some information
unique to the sender.
The digital signature must be easy to generate.
The storage of a copy of the digital signature must be simple.
Forging the signature must be computationally infeasible, i.e., either by
constructing a fraudulent signature for a given message, or constructing
a new message with an existing signature.
The signature must be easy to recognize and verify.
Secret Key Signatures
This approach involves a central authority that is trusted by everybody. Each
user shares his/her secret key with the CA.
Alice wants to send a signed plaintext to Bob. She generates the string (B,
RA, t, P) and encrypts it with her secret key KA. This, along with her identity,
is sent to the CA as message 1.
The CA, on observing the message from Alice, decrypts it with her key
KA and extracts the plaintext P, time-stamp t and the random number RA. CA
then combines these strings and signs it with its own signature KCA. This
encryption, along with A, RA, t and P, is again scrambled using Bob’s secret
key to form the message 2, and this is sent to Bob.
Bob decrypts it with his secret key, KB to extract P and KCA(A, t, P). The
signed message from CA is stored by Bob as a proof that Alice had sent P to
Bob. In case of any dispute, when Bob claims to have received the message
from Alice and she denies it, the CA can decrypt the KCA(A, t, P) portion of
the message received by Bob and verify the fact that the message was indeed
sent by Alice to Bob.
SHTTP
SHTTP (Secure HTTP) is a scheme proposed by CommerceNet, a coalition
of businesses interested in developing the internet for commercial uses.
Current HTTP implementations only provide modest support for the security
mechanisms necessary for commerce. SHTTP provides a wide variety of
mechanisms to provide for confidentiality, authentication, and integrity to
HTTP clients and servers. Separation of policy from mechanism was an
explicit goal in the design of this protocol. The system is not tied to any
particular cryptographic system, key infrastructure, or cryptographic format.
Secure HTTP is a secure message-oriented communications protocol,
designed for use in conjunction with HTTP. It is a superset of HTTP, which
allows messages to be encapsulated in various ways. Encapsulations can
include encryption, signing, or message authentication code (MAC) based
authentication. This encapsulation can be recursive, and a message can have
several security transformations applied to it. SHTTP also includes header
definitions to provide key transfer, certificate transfer, and similar
administrative functions. SHTTP appears to be extremely flexible in what it
will allow the programmer to do. SHTTP also offers the potential for
substantial user involvement in, and oversight of, authentication and
encryption activities.
The protocol provides symmetric capabilities to both the client and the
server (such that, equal treatment is given to both requests and replies, as well
as for the preferences of both parties) while preserving the transaction model
and implementation characteristics of HTTP. Several cryptographic message
format standards may be incorporated into SHTTP clients and servers.
SHTTP supports interoperation among a variety of implementations, and is
compatible with HTTP. SHTTP aware clients can communicate with SHTTP
oblivious servers and vice-versa, although such transactions obviously would
not use SHTTP security features.
SHTTP does not require client-side public key certificates (or public
keys), as it supports symmetric key-only operation modes. This is significant
because it means that spontaneous private transactions can occur without
requiring individual users to have an established public key. While SHTTP is
able to take advantage of ubiquitous certification infrastructures, its
deployment does not require it.
SHTTP supports end-to-end secure transactions, in contrast with the
original HTTP authorization mechanisms, which require the client to attempt
access and be denied before the security mechanism is employed. Clients
may be “primed” to initiate a secure transaction (typically using information
supplied in message headers); this may be used to support encryption of fill-
out forms, for example. With SHTTP, no sensitive data need ever be sent
over the network in the clear. SHTTP provides full flexibility of
cryptographic algorithms, modes and parameters. Option negotiation is used
to allow clients and servers to agree on transaction modes (e.g., should the
request be signed or encrypted or both, and similarly for the reply)
cryptographic algorithms (RSA vs DSA for signing etc.), and certificate
selection.
SHEN
SHEN is a scheme proposed by Phillip Hallam-Baker of CERN. Like SHTTP
it is a high level replacement for the existing HTTP protocol.
SHEN provides for three separate security-related mechanisms:
1. Weak authentication with low maintenance overheads, and without
patent or export restrictions.
A user identity must be established as genuine. Unauthorized access
must be improbable, but security from all possible forms of attack
events need not be provided.
2. Strong authentication via public key exchange.
A user identity must be established as genuine. Unauthorized access
must be impossible except by random chance, or by access to unknown
technology.
3. Strong encryption of message content.
The data must not be transmitted in a form comprehensible to a third
party; with an identified party acting as guarantor in this respect.
Although SHEN has existed as a proposal for nearly two years, no
browser or server vendor has implemented it.
CONCLUSION
The information superhighway has seen exponential growth over the past few
years. Society is becoming increasingly reliant on informational, rather than
physical, transactions. The electronic medium is replacing the physical
medium. The expected total volume of trade to be carried out over the web is
growing at an exponential rate. This makes the information available on the
network a valuable commodity, and raises numerous questions about its
security, access control, privacy, authenticity of communications, and
unforgeability of the data.
The internet today is a vast frontier of unknown elements, including new
types of software, new discoveries of security flaws, and unfriendly
neighbors. Electronic commerce and information security are growing areas
of concern to user communities. New applications, new users, and faster
connections have spurred the internet to become an important medium for
communication, information dissemination, and commerce. As the internet
becomes the basis for electronic commerce and as more businesses automate
their data-processing operations, the potential for unauthorized disclosure of
sensitive data increases. Online databases are becoming increasingly large
and complex. Sensitive data is transmitted on communication lines, and often
stored offline. As a result, the efficient, economical protection of enterprise-
critical information has become increasingly important in many diverse
application environments. Nevertheless, planned and current security policy
regarding the internet is not well developed. The most secure technical
solution to preventing attacks launched from the internet is to unplug the
network from the computer. This solution is not viable in today’s business
climate. Instead, the components that comprise e-commerce systems must be
adequately secured.
Securing e-commerce must occur on four fronts: (1) securing the web
clients, (2) securing the data transaction, (3) securing the web server, and (4)
securing the network server operating system. The security of e-commerce
systems, though, is only as strong as their weakest component. A failure to
secure any one of these four components of electronic commerce may result
in the entire system being insecure. Organizations need to be proactive in
fortifying their resources linked to the network. It is quite reasonable to
tolerate a flaw that is rarely exposed and assume that having occurred once it
is not like to occur again. It is also reasonable to assume that logically
independent failures will be statistically independent, and not happen in
concert. In contrast, a security vulnerability, once discovered, will be rapidly
disseminated among the growing community of hackers, and will be
exploited on a regular basis until it is fixed.
Security remains the biggest obstacle in many individuals and
organizations reposing full faith in the web. It is a major issue facing
organizations today. We live in an era characterized by complex computer
environments, by multiple computer platforms, and by vast conglomerates of
integrated computer networks. As technology advances and ushers in new
innovations in network communications, new loopholes will be discovered
which can compromise the security of the systems. Implementing security
across the entire enterprise can, therefore, be a perplexing and overwhelming
task. The crux of the matter is that network security is not a static subject.
Internet development has been dynamic and so will be security issues it. In
the future though, more proven tools and techniques will be available to
combat internet crime. But at the same time the gravity and scale of
electronic crimes may also increase. The future of the internet is an exciting
prospect and does hold many surprises.
SUMMARY
The distributed nature of electronic commerce requires information flow
among various entities such as buyers, sellers, and intermediaries. The
technological infrastructure that can assure secure message transfer between
interacting entities is essential for the growth of the electronic commerce.
This chapter discusses the threats and requirements for creating a trustworthy
transaction environment. Cryptography plays a fundamental and essential
role in enabling such a transaction environment. In this chapter, the basics of
cryptography, cryptanalysis, conventional encryption models, and public key
cryptosystems are described. Standard cryptographic algorithms such as
DES, triple DES, IDEA, RSA, MD5, and SHA, along with the vulnerabilities
have been discussed here. These algorithms are used for establishing a
transaction environment a that supports authentication, integrity,
confidentiality, and non-repudiation. In addition to these algorithms,
Kerberos as an authentication mechanism is also described. Further, the
chapter discusses digital signatures, public key infrastructure to support
digital certificates, and the role of certification and registration authorities.
Finally, the chapter deals with commonly used protocol implementations, for
enabling secure web commerce, such as SSL, SET, and SHEN.
APPENDIX
REVIEW QUESTIONS
1. What is meant by integrity of a message? Describe a technique to ensure
the integrity of an e-mail message.
2. What is a digital certificate? Describe the commonly used standard for
the digital certificate.
3. Describe a symmetric key cryptosystem? What are the important issues
related to key distribution and management.
4. What is a public key cryptosystem?
5. What is Public Key Infrastructure (PKI)? Describe the role of the
certification authority and the registration authority?
6. What is digital signature?
7. Describe a technique used for the non-repudiation of an electronic
commerce transaction?
8. What is a low encryption attack on the RSA algorithm?
9. What is secure hash algorithm? Compare it with the Message Digest,
version 5.
10. Briefly describe the secure electronic transaction (SET) protocol.
11. Compare Secure Socket Layer (SSL) and Secure Electronic Transaction
(SET) protocols.
EXHIBIT 1
Electronic Mail Policies
1. All current students, faculty and staff will have an e-mail account. E-
mail address directories are made available for public access.
2. Anonymous re-mailer software cannot be installed. The faculty, students
or staff cannot use anonymous re-mailers for any purpose.
3. The e-mail system will provide a single, externally accessible e-mail
address for faculty, students, and staff. The address will not contain the
name of internal systems or groups.
4. Both primary and secondary mail servers will be inside the firewall. All
messages will be scanned for viruses and other maligned content, by a
gateway based anti-virus software.
5. Users will be able retrieve e-mail through IMAP (Internet Message
Access Protocol) or POP3 (Post Office Protocol) services, from inside
the network. From outside the network, users will be allowed to access
their mail only the using “webmail” service, available through the IIML
web page. Authentication is enforced for retrieving messages.
6. E-mail servers will be configured to refuse relaying any e-mail
addressed to non-IIML domains.
7. IIML is not responsible for the retention of the e-mail messages. The
users are responsible for proper backup and archival of their respective
e-mail messages.
8. A content analyser will be installed at the gateway and configured by the
designated vendor to monitor any abusive content in the messages and
attachments. The content analysis will be done both for messages
originating from the internal network and for those from outside
networks.
World Wide Web Policies
1. The Institute web server (www.iiml.ac.in) will be placed inside the
firewall, in the De-Militarized Zone (DMZ). All the other web servers
will be hosted in the internal network. All HTTP requests from outside
to internal web servers will be processed through the firewall and
appropriate reverse proxy servers.
2. All files downloaded over WWW will be scanned for viruses or other
malign content, using a gateway based anti-virus software and content
analyser.
3. All web browsers will be configured to use Hypertext Transfer Protocol
(HTTP) proxy.
4. No offensive or harassing material should be made available via the
IIML web site. Periodic checks will be done on all public and private
web pages by the web administrator and any undesired material will be
immediately removed.
5. Users are responsible for posting personal and other valuable
information through forms. Users shall use the secure form feature to
encrypt information posted through forms.
6. No personal commercial advertising should be made available via the
IIML web site.
7. Users are permitted to have their personal web sites at designated
locations on the web servers. The users are responsible for the content
and backup of their web pages.
8. A local archive of web authoring tools will be maintained and made
available for internal use.
9. The web server software and the software of the underlying operating
system will be updated periodically, with appropriate batches and
updates by the WWW administrator.
An excellent source for security policy formulation is Internet Security
Policy: A Technical Guide, published by the National Institute of Standards
and Technology, and can be found at NIST web site: http://csrc.nist.
gov/isptg/
_____________________________
Reprinted from a case study by Prof. V. Sridhar, Dilip Mohapatra and Mr. P.
Mohanan (IIML), and Ravikiran Bhandari and Keshava Murthy D.G.
(Banglore Labs.) Voice and Data, 8, No. 8 (February 2002).
Learning Objectives
This chapter covers the following topics:
1. What is supply chain?
2. Importance of supply chain management
3. Role of Information in supply chain management
4. Impact of electronic commerce technologies on supply chain
management
a. Impact on procurement
b. Impact of production planning and Inventory
c. Impact on distribution
As seen in the previous chapter on Electronic Data Interchange (EDI), the use
of Information Technology can have a profound impact on the procurement
process. The transition to EDI-based procurement requires the use of standard
document formats for expedited processing of documents, exchanged
amongst trade partners over the information network. The adoption of EDI
based procurement reduces the lead-time, improves the supplier’s
coordination and expands market reach. These benefits accrue to companies
that are able to adopt the EDI standards and roll-out EDI based procurement
setup. The arrival of web-based electronic commerce with its ubiquitous,
easy-to-use interface has created an opportunity of streamlining the
information flow in the entire procurement process. The opportunity can be
exploited in streamlining both the upstream and the downstream of supply
chains.
For the organization involved in offering products or services to a
customer, before products end up at the hands of end-consumer, there is a
sequence of activities involving the basic procurement and supply of raw
materials, storages and warehouses, assembly, manufacturing, processing,
distribution and retail. The typical manufacturing/assembly and service
supply chains are shown in Figs 10.1 and 10.2 respectively.
Impact on Distribution
In an organization, the role of the distribution network is of paramount
importance. It is through the strength of the distribution network that firms
are able to reach a large and geographically dispersed customer base and
ensure the delivery on demand by appropriately stocking the product in close
vicinity. The innovation in the distribution network can offer a huge
competitive edge, by optimizing the inventory levels that has to be
maintained in nearby warehouses to support the demands generated by end-
consumers and aggregated and filled by intermediaries.
Intermediaries, in a communicators’ role, serve a prominent role as a
collector of information on consumer preferences, demand trends and
demand sensing, and also disseminate the new product information to the
consumers. Also, they disseminate the price, quality, functionality and
availability information about the existing products to their customers. From
the consumer’s viewpoint, these intermediaries help reduce the search costs.
The intermediaries, in the logistics support role, facilitate the movement of
physical goods from manufacturers to the end consumers, share the risks of
inventory management and distribution and through aggregation of demand
information, create scenarios for exploiting economies of scale in
transportation and distribution of goods. The intermediaries also assist in
providing customer service and bear the costs and risk of being the front line
of contact with the consumer.
Thus, the manufacturer and distributers operate in a principal-agent
relationship. While doing so, the agents not only add value through their
services, but also add to the cost of product. The end consumers’ price of the
product consists of the following:
1. Search cost
2. Production cost
3. Coordination/Distribution cost
4. Profit
As we saw in the above two sections, electronic commerce reduces the
search cost. Manufacturing costs are also impacted by the use of the Internet
and e-commerce technologies in procurement and production planning,
scheduling and inventory management. A lot depends upon the nature of the
product. In an extreme case, it can transform the manufacturing operation to a
“virtually integrated” collaborative demand-driven supply network operation,
leading to competitive and efficient operation. The coordination cost consists
of distribution logistics and intermediaries costs. As stated earlier, with the
ubiquity of the Internet and electronic commerce technology, the role of
intermediaries involved in the distribution channel has come under severe
restructuring, especially in the area of the software products, music, movie
rental and book distribution. Many traditional powerhouses like Tower
Records, Blockbuster and Barnes and Nobles have seen a drastic impact,
much to their peril.
The major impact of electronic commerce on the distribution channels
emanates mainly from its ability to provide a wider outreach and enhance
communication capability. The electronic commerce technology enables and
enhances the capability of collecting and processing information at a lower
cost. As the collected information is digital in nature, it can be processed in
negligible time and cost. The electronic commerce technology platform
establishes an efficient two-way channel of communication and information
processing; enabling and potentially leading to integration of disparate
processes of information collection, manipulation and communication.
Consequently, it creates an integrated collaborative platform where the
information collected at various points can be made accessible. Through its
communication ability and easy to use web-based interface, it offers a
capability to create a marketplace where a large numbers of buyers can
interact and transact with a large number of shares and vice versa.
The impact of electronic commerce on the role of intermediaries largely
depends upon the nature of goods under consideration. In order to see the
impact, we can classify goods in the following three categories:
1. Digital Goods and Digitally Deliverable Services
2. Physical Goods
Digital Goods and Digitally Deliverable Services
In an electronic commerce environment, digital goods refer to all such items
that are created, stored, distributed and finally used in electronic form.
Packaged software, mp3 music, DVDs, e-books and online games are some
of the prominent examples of digital goods. Since these goods are always in
electronic format, with the proliferation of the Internet and electronic
commerce platform, the delivery of these goods can be made over the
Internet instantly. The traditional commerce has been geared for physical
distribution and delivery of goods. In such an environment, there was no
choice but to store these goods in physical media such as floppy disks and
CDs and package them in box like all other physical goods to facilitate the
use of existing distribution channels. Once received by the buyer, either
through a retailer or some other traditional channel such as mail
order/catalogue store, the user installed/loaded the product on a compatible
electronic device and made use of the product. The physical media used for
delivery had no further use. The model in the process incurred production
costs for packaged software, music CD etc. Further it incurred the
coordination cost as it traversed through the traditional channel. Since the
digital goods are created, stored and used in electronic format, the
proliferation on the Internet through electronic commerce platforms offers an
opportunity to deliver these goods over the network in real-time to
consumers’ devices. In this case, there is no need for production and
distribution through traditional channels. The electronic commerce server
platform can be used for storing and displaying these goods. The buyers can
directly transact with the producer of the digital goods and have them
delivered online. Alternatively, digital goods can be displayed and transacted
through a value added intermediary and delivered directly to the customer. In
either of the cases, the product resides and is delivered in electronic format
from master servers. Thus, the costs associated with logistics issues and
stocking of the product are almost eliminated. Figure 10.6 shows a simplified
view of distribution chain for the digital goods.
The Background
Hindustan Unilever Limited (HUL) is the largest FMCG Company in India
with a turnover of ` 110 bn. It operates one of the largest distribution
systems in the world. The sales and distribution system services one million
retail outlets directly through a network of 7000 stockists and 50 depots
across India.
RS Net is the Internet based system connecting Hindustan Unilever to its
Redistribution Stockists (RS).
The Business Case
While Hindustan Unilever has a large, successful and dominant distribution
system, the battle for growth and outlet leadership required a fundamental
shift in selling systems. The specific thrusts were:
Replenishment driven Primary Sales (Sales from Hindustan Unilever
to the Stockists),
Focus on Secondary Sales (Sales from the Stockists to the retailer),
aided by online availability of information, and
Enhance communication and build the customer management
community across the geographical breadth of India.
The business aimed to release inventory, release field force time by 50
percent, ensure full line availability at the retail outlet and thereby achieve
growth. Embracing a new generation of Information Technology and
specifically the Internet could only do this.
An E-Commerce team was constituted comprising of IT and business
managers in early 2001 to achieve this.
The IT Objective
The IT objective was clearly to achieve connectivity with the stockists.
The stockist is an independent businessman who is closely associated
with the company, but not under its executive control. When the project
started more than one-third of the stockists did not own a computer and had
not seen one in their lives. The Internet was a far away dream. Even the
two-thirds who had a computer had a plethora of packages, ranging from
spreadsheets to DOS based packages to mini ERPs. No standardization of
formats or product/entity codes existed.
The project started in the beginning of 2001. The target was to connect
stockists comprising 80 percent of turnover by 2002.
What Has Been Achieved – IT
The project team set out to achieve the following:
Computerize every stockist.
Connect the stockist to the Internet.
Migrate stockists on disparate systems to a set of about 30 “approved”
local packages with whom interoperability would be achieved. One of
the best decisions the team took was NOT to migrate all stockists to a
single package, which would have been unachievable.
Build Interfaces with each of the 30 “approved” stockist packages.
Automatically upload daily sales, stock and market information
through the Internet from every connected stockist.
Compute a replenishment-based order and offer it back to the stockist
for confirmation.
Communicate the confirmed order to one of the 50 depots, where a
stand-alone version of the ERP – MFG/Pro is run, from where it would
be serviced.
Provide intelligent secondary sales information via the Internet to the
sales team wherever they are.
Deals were struck with Internet Service Providers (ISP) to extend the
Internet access to the places that were to be connected. Infrastructure being
unreliable in interior India, the whole application was designed so that not
more than five minutes of the Internet connectivity was required per day.
About one million individual product codes were mapped at stockist points
to achieve standardization of information. Every stockist was trained to
access the Internet and use the system. In a number of towns, RS Net
represented the first foray of the Internet into the town!
At the back end, the infrastructure had to be robust with 99.5 percent
uptime and scalable to handle 250,000 orders and 60 million records per
annum. The application was developed on a UNIX platform with a J2EE
based 3-tier architecture using iPlanet Web Server, Weblogic Application
Server and Oracle Database Server. For content management, Vignette
content management server was used. Clearly high levels of security were
required and built into the application and the infrastructure.
What Has Been Achieved – Business
RS Net provides linkages with the stockists’ own transaction systems,
enables monitoring of stocks and secondary sales and optimizes orders and
inventories on a daily basis. Information on secondary sales from all across
the country is now available on RS Net every day. Also, the stock service
levels at stockists can be monitored on a daily basis.
Riding on RS Net, the business has shifted its focus on secondary sales
in the connected stockists. A large amount of inventory has been released
from the stockists. Sales force time in the market has gone up and so has the
number of lines sold. The stockists’ role has changed from that of an
investor in stocks to that of a service provider.
With RS Net, Hindustan Uniliver’s sales and distribution system has
been relaunched with fundamental business process changes to remain a
source of competitive advantage and to deliver profitable growth.
SUMMARY
Ubiquity of the Internet and web based electronic commerce platform have
greatly influenced supply chain management. The farsighted companies have
recognized the information sharing, two-way communication ability of the
electronic commerce platform and how it has impacted the information
asymmetry. This instant information sharing amongst all partners of the
chain, further accentuated by the communication ability, has created far more
efficient procurement options and has lead to the development of alternate
sources for supplies. The role of information in countering the bullwhip
effect has been widely recognized. Imperfect information and its
amplification leads to inventory build up at every stage. The collaboration
and information sharing capability plays an important role in mitigating the
bullwhip effect and also leads to lowering the inventories at every stage of
the supply chain. The major impact of electronic commerce can be seen in
facilitating the emergence of demand driven manufacturing leading to the
formation of demand driven supply network (DDSN), as seen in the case of
Dell, where the component suppliers deliver the components directly to the
assembly lines, based upon the production schedule generated, by taking in
account demand generated by all the customized orders, every two hours.
Finally, the traditional distribution channel consisting of intermediaries, who
facilitated the physical movement of goods and information related to
demand, customer preferences, feedback, and payment in both directions,
have seen the widespread impact. The information sharing and
communication ability of electronic commerce has made the role of many
intermediaries redundant. This has lead to the restructuring of the supply
chain, depending upon the nature of the product. The electronic commerce
has been able to create a huge impact due to restructuring of the supply and
distribution chains in the digital products, services, branded goods and
standardized products even with low volume and high cost. Manufacturers,
like Dell, have successfully created a competitive direct to customer model
using the electronic commerce platform. In the low priced and high volume
category of product manufacturers, like Hindustan Unilever, electronic
commerce technology platform has enabled them to eliminate some of the
intermediaries and thus, reduce the channel length and in turn, the
friction/coordination cost.
REVIEW QUESTIONS
1. Define supply chain management and how it relates to business
competitiveness?
2. What is bullwhip effect and how information sharing can be used to
manage its impact?
3. What is the impact of the following on supply chain management?
a. Increasing Globalization
b. Outsourcing
c. Internet enabled E-Commerce
4. What is the role of electronic commerce in promoting disintermediation?
5. Discuss the role of B2B electronic commerce exchange in altering the
supply chain?
6. What do you understand by demand driven manufacturing?
7. What are the key enablers for implementing the demand driven
manufacturing?
8. Describe the minimal set of technological platform necessary for
moving to demand driven suppliers’ network.
9. Describe the characteristics of the products where manufacturers can
directly sell to consumers? Give an example?
__________________________
1Kavitha Rao prepared this case as a basis for class discussion rather than to
illustrate either effective or ineffective handling of an administrative
situation.
Learning Objectives
This chapter covers the following topics:
1. Introduction to Payment Systems
2. Basic Characteristics of Online Payment Systems
3. Prepaid Electronic Payment Systems
4. Post-Paid Electronic Payment Systems
5. Comparison of some existing based on requirements Payment Systems
Economy Issues
In order to become an accepted economical instrument, a digital payment
system needs to provide a trusted, reliable and economically feasible service
to a sufficiently large user community.
Operational: A system should be deployable immediately, i.e., the
testing of the payment system should not be so protracted as to render
the mass use impossible.
Large User Base: The payment system should be used by a large
number of customers. The size of customer base willing to use the
digital payment system affects the merchant’s attraction to it, while
currency acceptance by large number of merchants affects the size of
user base.
Low Risk: The electronic payment system should minimize the risk of
financial loss associated with the use of such payments systems, it
should at best be limited and controlled. In order to develop trust, users
should be protected, to some extent by the payment system, from the
financial losses emanating from system misuse.
Reliability: An electronic payment system must be highly reliable in its
operation. It should ensure high availability as even a temporary failure
can cause uncontrollable losses to its user base.
Conservation: It refers to the conservation of value stored in digital
currency, over a period of time. It should be easy to store and retrieve
the value. The value of money should be lasting in nature, it should
diminish when spent, rather than become invalid with the passage of
time.
Ease of Integration: The electronic payment system needs to be
integrated with applications that conduct the electronic commerce
process over the network. The process of integrating electronic payment
systems with e-commerce applications should be easy, to facilitate their
growth of usage.
The following table compares the various payment systems for the economic
issues described above.
Ease of Use
The usability of the electronic payment system plays an important role in its
being adopted by the user community. The electronic payment system should
be easy for the user to relate to, accessible, and simple enough to understand.
It should operate in a fashion that builds confidence in users. At no stage
should the users feel lost or confused in the process of making payments.
Unobtrusiveness: This refers to the operational transparency of the
electronic payment system. A payment process should be clear, concise,
simple to understand, and yet should operate with minimal interruption
and intervention from the user.
Low Latency: The payment protocol used in the transaction should have
a low performance overhead. It should not become an overhead on the
purchase transaction.
Low Transaction Costs: The overhead costs charged to the users, in
making the payment through the electronic payment system, should be
extremely low and depend on the value of the transaction. It acquires
added significance in the case of micro payments.
Hardware Independence: Users should not require specialized hardware
to make use of the payment system. Hardware dependence, which is
expensive, would vastly limit the popularity and hence the use of the
payment system itself.
Based on the requirements discussed above, a comparison matrix is
presented belows:
SUMMARY
From the barter system, the payment mechanism has evolved to being
notational in nature. In the notational systems the value is stored with a
trusted third party (such as bank), and we transact using notational
instruments such as cheques. It has further evolved into a credit system which
permits transaction without any stored value with the trusted entity. To
facilitate transactions in the emerging electronic commerce environment
online payment mechanisms have become the need of the time. As a result,
several forms of online payment mechanisms have been proposed and
implemented. For the wider acceptance and viability of these mechanisms,
they are expected to exhibit certain characteristics. In this chapter, these basic
characteristics of online payment systems have been described. The newly
emerged online payment systems can be classified as prepaid and post-paid
payment systems. The basic operations of some of the prepaid payment
systems and post paid payment systems have been illustrated. The prepaid
payment systems illustrated in the chapter include eCash, CAFÉ, Mondex,
MilliCent, MicroMint, Netbill, Minipay and Netfare. The post-paid payment
systems described in the chapter include iKP, CyberCash, SET, FSTC,
Mandate and NetCheque. Finally, the chapter presents a comparison of these
payment systems with regards to basic requirements of an online payment
system.
REVIEW QUESTIONS
1. Discuss the basic requirements of an online payment system.
2. What are micro payments? What are the special considerations involved
in the design of an online micro payment system?
3. Discuss and differentiate between prepaid and post-paid electronic
payment systems.
4. Describe an online payment transaction in the Mondex Smart card
system.
5. Describe a transaction in the FSTC payment system.
6. Define interoperability in the context of online payment systems.
7. What are various security issues in the context of online payment
systems?
8. Describe what is meant by scalability, in an online payment system.
Indian Railways, one of the largest railway networks in the world, has
recently introduced the facility to book tickets through, the Internet. These
tickets are subsequently bonce delivered the Indian Railways Catering and
Tourism Corporation (IRCTC) is currently handling Internet ticket
reservation in collaboration with a number of banks, which have established
payment gateways on the IRCTC site to facilitate payment of money online.
In order to tap online customers, the State Bank of India, India’s largest
bank, has introduced its e-Rail system in 2003. With the help of e-Rail,
customers can now pay for their tickets directly from their bank account. In
this report, we have sought to study the various aspects of the e-Rail system
—its business model, the payment and security mechanisms employed, etc.
Background
The State Bank of India is the largest public sector bank in India in terms of
profits, assets, deposits, branches, and employees.
The origins of State Bank of India date back to 1806, when the Bank of
Calcutta (later called the Bank of Bengal) was established. In 1921, the Bank
of Bengal and two other banks (Bank of Madras and Bank of Bombay) were
amalgamated to from the Imperial Bank of India. In 1955, the controlling
interests of the Imperial Bank of India were acquired by the Reserve Bank of
India and the State Bank of India was created by an act of the Parliament to
succeed the Imperial Bank of India.
The bank has undergone large scale changes in the last decade as it began
to modernize its operations and processes. It has the largest network of ATMs
and has also introduced several value-added features for its customers.
Information Technology Usage in SBI
The bank is pursuing an aggressive IT policy with the objective of achieving
efficiency in internal operations and of meeting customer and market
expectations. To carry this strategy forward, several IT projects have been
launched.
1. Universal Computerization Project: Computerization of all the
branches of the SBI group under the Universal Computerization Project
(UCP) on LAN-based Bankmaster software was completed with the
computerization of the remaining 7,526 branches between May 2003
and January 2004.
2. ATM Project: ATMs are the most dynamic retail channel today in
terms of the transformation they are bringing about in banking habits
and their popularity with customers and branch staff alike. Various
initiatives that have been taken on this front include:
• Installation of 2,247 ATMs during the financial year 2003–04, taking
the aggregate number to 3,814 ATMs, covering 1,152 centres.
• All stand-alone ATMs moved to the networked platform.
• Creation of a single ATM network across banks of the State Bank
Group.
• Strengthening of the ATM project through change in strategy,
improved marketing, and customer/user education.
• The current ATM card base stands at 5.8 million. The turnaround time
for ATM cards issue has been brought down to 7 to 10 days.
• Entering into bilateral tie-ups with other banks like HDFC Bank and
UTI Bank for sharing of ATM networks.
• Creation of ATMs as cash points for SBI Cardholders.
• Enabling of value-added facilities such as payment of premium on SBI
Life policies, payments on account of SBI Credit Cards, and fees of
certain schools and colleges and mobile topping up through ATMs.
• On-line booking of Railway season ticketing provided through ATM at
CST station, Mumbai. The Group’s network of ATMs has also been
opened to credit card associations (both Visa and Mastercard).
3. Core Banking Project: The Bank is moving towards a centralized
database with a state-of the-art core banking solution with capability for
on-line real-time transaction processing. After the first pilot branch went
live in August 2003, 40 more branches have gone live by March 2004.
Utilization of core banking solution to create new and innovative
products coupled with efficient transaction handling through centralized
processing will help the bank in delivering value-added services to
customers.
4. Trade Finance Project: The bank has identified anintegrated trade
finance solution, the ‘Exim Bills’ software, which has been customized
for Bank’s operations, will allow the bank to transact its foreign as well
as inland trade finance and bills business on the same central database.
Compared to the software currently in use, the new software brings in
an improvement in terms of efficiency, reduced operating costs, and
accurate and timely MIS support. After the pilot project in August 2003,
it has been rolled out to 94 large and trade finance intensive branches by
March 2004. The bank has acquired an Internet access module for trade
finance software which will enable commercial customers to put in
requests for various transactions directly from their offices through the
Internet.
5. Internet Banking: The channel is an extremely comprehensive product
for both retail and corporate use. It has acquired real-time transaction
processing capability and has been supporting the business initiatives of
the Bank in the areas of utility bill payments, IIT application money
receipts, railway ticket bookings, credit card payments, insurance
premium payments etc. More products are added regularly to meet
customer demand. Corporate Internet banking provides customised
products to large corporates. As on the 31st March 2004, 1,110 branches
provided Internet banking service covering over 300 centres. The
facility has so far reached 2,45,000 customers under the retail segment
and 7,800 customers under the corporate segment. Request for opening
of Internet banking accounts through ATMs has been operationalized for
card holders.
6. SBI Connect: SBI Connect, a wide area network (WAN) and a crucial
infrastructure platform, will make real-time transactions between
branches possible. Critical applications such as core banking, ATM, and
Internet banking depend on the WAN for successful functioning. As on
March 31, 2004, 4,215 offices of the SBI Group (2,619 offices of SBI
and 1,596 offices of Associate Banks), covering over 300 centres, were
connected through the SBI Connect network.
7. Other Projects: The bank’s telebanking service provides certain
selected banking activities like enquiry on the bank’s products, foreign
exchange rates, issue of draft or cheque book and delivery thereof at the
customer’s residence, statement of account by fax, etc. SBI Homepage,
the bank’s website, provides a wide range of information and is also
now available in Hindi. It has been redesigned to provide a wide range
of information to the Bank’s customers. The bank has a well laid system
for prompt redressal grievances customers. The SBI helpline
established, at all LHO centers, are equipped with toll-free telephone
lines, fax, and e-mail for providing quick and complete information on
the bank’s products and services and to enable customers to have their
grievances redressed promptly.
Introduction of single window services at more than 10,000 branches
across the SBI Group has enabled customers to conduct their
transactions efficiently. The Bank is aiming to introducing the single
window system at all its branches.
SBI pioneered in introducing Online Tax Accounting Software
(OLTAS) for collection and transmission of corporate tax collection data
to RBI, Nagpur, on T+1 basis for the convenience of the government, its
customers, and to counter the growing competition for government
business. As on end-March 2004, 3,466 branches were OLTAS-enabled.
As a measure of providing value-addition and convenience to the bank’s
customers in the conduct of government transactions, the single window
service for government transactions has been extended to over 2,800
branches.
Management Information System (MIS): MIS in the bank is being
constantly upgraded to cater to the constantly growing appetite for
information for decision support, for innovating customized products,
and for statutory needs. Full computerization of the branch network has
made a significant contribution in enhancing data quality, reliability, and
timely availability.
Credit Information System (CIS): A solution developed in-house for
meeting the information needs on the loans and advances portfolio in its
entirety, has been implemented at all branches in the Bank, dispensing
with the need for compiling reports manually at the branches. The
solution has also resulted in the development of a rich database which
will enhance capabilities in the areas of product development, cross
selling, risk management, and business intelligence in the days ahead.
CIS together with other MIS applications are in the process of being
rolled out to cover the entire Group.
Branch Interconnectivity at SBI
Need For Connectivity Between Branches
1. Facilitates ‘anywhere anytime’ operation of bank accounts (associated
with any branch) by the account holder. Almost all banking transactions
like cash withdrawal, issue of cheque book, demand draft and account
statement can be done in this mode of anywhere anytime banking.
2. The bank’s daily routine internal work like interest application,
balancing, tallying cash, day book writing, posting in ledgers/accounts
reconciliation of inter-office accounts/DD purchases (withdrawal of
money from out station accounts), mail transfers (sending money from
one account at one place to another account at the same place or other
place etc.), earlier used to take almost half a day or even more. With
computerization between branches, these activities are done
instantaneously by computers. As a result, there is an increase in the
effective number of business hours available in a single day.
3. Interconnectivity between branches facilitates easy and fast reporting (of
data/banking transactions/feedback) from innumerable branches (+
9000) to a single point of controllers, which is otherwise a mammoth
task consuming all the available time of a large number of employees.
4. Easy and fast retrieval of data in various forms is ensured by the
controllers from a large number of branches for important
managerial/executive functions like planning, policy, strategy making,
and identification of problems.
5. Computerization is also facilitating outsourcing of innumerable daily
routine functions that are not directly connected with customers (for
example, payment of salary and allowance, carrying out standard
instructions of customers, like payment of school fee, feeding various
types of data into the system, editing existing data etc.)
6. Computerization has helped in the introduction of the single window
system (all banking transactions being done through a single counter)
and handling increased number of products.
7. Frauds can effectively be checked as system and procedural errors can
easily be noticed without scope for suppression or distortion. Recovery
of bank dues can effectively be followed up with the information readily
available at multiple points.
8. Interconnectivity between branches facilitates multiple channel Interface
(ATM, Internet banking, tele banking).
Security Threats
The different types of security threats possible are in the form of virus from
Internet sources, hacking, theft, manipulation of data, and intrusion of
privacy.
Security Solutions Available Systems and procedures allow only authorized
people to have access during restricted hours.
The use of firewalls alone might not be a sufficient security measure against
the type of security threats listed above.
‘Verisign’ certifies the bank’s site and provides tools to overcome security
threats. It facilitates data transfer in an encrypted form so that data integrity
can be maintained.
SSLs (Security Socket Lock) of 34Bit and 62Bit are currently being used.
E-RAIL: Overview
The e-rail model is a site-to-site integration system where the two
websites (www.irctc.co.in and www.onlinesbi.com) talk to each other. The
user independently registers and explores inside the railways website and
then identifies his requirements. The IRCTC system passes a query to the
passenger reservation system to check for fares and availability. Upon
receiving this information, the customer opts to pay by debiting his account
with SBI, with the help of his credit card number. Only after this does the site
www.irctc.co.in send a string to the site www.onlinesbi.com. Then the
customer is taken to www.onlinesbi.com and there he enters his user ID and
password into this site and is then subjected to the IP protocols security
cover (i.e., 128bit SSL security). After he clicks on the Confirm message his
account is debited and he logs out of the SBI system. Thus, the entire stage
for the financial transaction is secured.
On the IRCTC front, the existing Railway Passenger Reservation System
(PRS) has all the data and does the booking. The Internet system piggybacks
on the PRS and provides an extra booking arm. No train or availability data is
stored in the Internet system. They are all retrieved from the PRS system,
live, against user queries. The booking is also done directly on the PRS
system.
Once this verification is carried out, it prints the physical ticket in Delhi
and they are sorted out on a city-wise basis and sent by courier.
These interactions can be depicted as shown in the diagram below2.
_________________________
1 This case was prepared by Jyotsna Jallepalli, Manoj Gaddam, Mrugendra
Shintre and Shreyas Gopi nath to from the basis of a class discussion rather
than to illustrate either the effective or ineffective handling of an
administrative situation.
2 SBI is not responsible for any delays or disputes regarding tickets.
Learning Objectives
This chapter covers the following topics:
1. Electronic Marketing
2. Influence of Electronic Marketing on a Product
3. Influence of Electronic Marketing on Physical Distribution
4. Influence of Electronic Marketing on Price
5. Influence of Electronic Marketing on Promotion
6. Influence of Electronic Marketing on Marketing Communication
7. Common Marketing Techniques
With the arrival of World Wide Web in the 1990s, the marketing element of
commerce was one of the first elements that saw a significant impact. Many a
leading technology-savvy company quickly discovered its potential to reach
the mass markets at negligible costs. In the early phase, it became a choice
platform for managing customer information, public relations, customer
service and support and sales under a single head. The impact of web did not
confine itself to the newly-anointed online marketing phenomenon, but it also
affected the offline marketing due to changes in the way of interaction of
buyers and sellers. In an attempt to understand to what extent electronic
commerce has influenced the marketing domain, the resulting benefits and
the emerging limitations, we have decided to follow the division made in the
following general set of activities, thereby studying the marketing mix in the
light of the evolution imposed by information technology.
Product
Physical Distribution
Pricing
Promotion
Marketing Communication
In addition to helping the reader to understand the specific changes that
the Internet has brought about in today’s economy, this division (or
marketing mix) seeks to point out the changes in marketing thought effected
by this technological phenomenon. These factors are important not only to
marketing managers but also to other functionaries in the company who need
to understand the complexity which the Internet added to business.
PRODUCT
The “product” part of the marketing mix represents the bundle of benefits that
is sold to organizations or consumers for money. These can represent either
tangibles such as physical consumer goods or services such as banking or
travel, or digital goods like software, etc. Internet commerce has transformed
many products from the brick-and-mortar economy to digital goods. Audio
music, videos, movies and even digital books are some of the examples of
transformed products. Electronic commerce has become a force of change in
many economies, with three effects on the marketing aspect of products:
New technology-based or technology-enhanced products have become
available.
Presentation, distribution and assortment of existing products has been
altered due to the availability of this new channel.
New opportunities for collaboration in business-to-business (B2B)
commerce have come to the forefront.
Each of these changes is due to the fact that online commerce has a very
different feel to it than face-to-face or telephone/mail order shopping.
What Kind of New Products does the Internet Create?
New products on the Internet are often differentiated by innovative uses of
technology. For instance, it is due to the innovative use of technology in the
form of putting Voice over Internet Protocol (VoIP) that we see the
emergence of new services such as Vonage, Skype, Google Talk or MSN
Messenger. Although, the initial use of some of these products/services was
powered forward by specialized technical users, yet, later a good high
number of them found common customers, who are relatively uninformed
about technology.
ILLUSTRATION 12.1 Skype
eBay acquired Skype for a whopping US $2.6 billion in 2005 in order to
capitalize on the growth of telecommunication traffic due to faster
convergence brought about by the Voice over Internet Protocol.
Skype delivers voice communication using an extremely cost-effective
business model, with almost complete elimination of expensive
infrastructure investments. The Skype service is built on top of the existing
data intensive network protocol IP. As of Februrary 2008, there were 240
million Skype customers. It has achieved the growth over a period of just
about three years, making it the fastest growing internet community ever.
Skype provides free telephonic calls to all the community members as
long as they have Skype installed on the other end of the line. A Skype user,
equipped with a PC internet connection, headphones and microphone can
experience the Skypes high voice quality. Skype also facilitates calls to the
standard fixed and mobile networks through the ‘SkypeOut’
function/service.
Skype is a peer-to-peer (P2P) Internet telephony software in which the
call is routed directly between computers of the two users instead of being
passed through a central server. The P2P service between the Skype users
take advantage of one existing Internet infrastructure, enabling Skype to
offer free unlimited phone calls to other Skype users globally.
Through the SkypeOut feature, users can make calls from PCs or PDAs
to fixed-line phones or mobile phones. The use of SkypeOut feature is on a
chargeable basis, where the charges vary from US $0.021 per minute (in
most of North America, Western Europe, Australia) to rates as high as US
$2 or higher, for calls made to some mobile phones and other territories like
the Dominican Republic. The Skype software supports various
heterogeneous platforms like Windows, Mac OS X and Linux. Skype can
connect up to five users in a conference-calls mode, irrespective of the
geographic location of the calls being made. The reliability of the service
can be gauged from the fact that CNN has carried out several of their video
conference interaction through the Skype service. The software, in addition
to the voice and video service, also supports file transfer across platforms
and instant messaging. The company offers PC to PC service, for free; thus
the main source of revenue is SkypeOut service. Additionally, phone sets,
PC headsets and related gadgets are the other source of revenue. The Skype
service, born on the network has witnessed a phenomenal growth in usage
and traffic volumes. The number of subscribers have almost trebled over a
period of two years. The subscription base of 95 million users at the first
quarter of 2006 has reached 276 million by the last quarter of 2007. The PC
to PC usage, offered free of cost, in terms of minutes has remained static
and has hovered around 6–7 billion minutes per quarter. SkypeOut minutes,
the revenue earning service, has almost doubled from 0.7 billion minutes
per quarter in the first quarter of 2006 to 1.7 billion minutes in the last
quarter of 2007, raising the revenue realisation from US $ 35 millions to US
$ 115 millions.
PHYSICAL DISTRIBUTION
The peculiarity of the Internet as a distribution channel is that it offers a
market place which is totally global and integrated from a geographic
standpoint. On the Internet, distance is no longer an issue.
In this distribution channel where physical location is of little importance,
companies that are successful are successful globally. Amazon.com, the
famous online bookstore as well as Virtual Vineyards, another successful
electronic business, distributes the shipments of their products to nearly 100
countries.
Despite the global characteristic of Internet-based distribution, it is
important to note that the primary language on the Internet is still English.
Many countries are reluctant to use English rather than their native language.
Very often, companies will translate their English website into the three or
four main languages spoken in the countries they are targeting (often
Spanish, German, French, Japanese, depending on the industry). For
example, the French software company Ubi Soft Entertainment sells online
and has web sites available in French, German, Spanish, Chinese, Japanese
and others. (http://www.ubisoft.com). This is the only way to reach the global
audience and to access people through the Internet distribution channel.
Companies which use the Internet as a main or additional distribution
channel should fully comprehend whom they are targeting and employ
relevant means to reach audiences sought, such as translating their sites into a
limited number of languages, from a cost-effective perspective.
The Internet has had a great impact on retail channel, changing the
processes, physical distribution and supply chain management. The Internet,
along with other IT systems (such as Just in time systems, EDI, and RFID)
has significantly altered the landscape of today’s retail universe. One of the
major influences of the Internet’s open standards for information flow has
been the Supply Chain Management (SCM).
According to Wikipedia, SCM is the process of planning, implementing,
and controlling the operations of the supply chain as efficiently as possible.
Supply Chain Management spans all movements and storage of raw
materials, work-in-process inventory, and finished goods from the point-of-
origin to the point-of-consumption.
Thus, managing the supply chain involves planning, organising and
optimising of one or more supply chain activities. It is about establishing a
long-term mutually beneficial partnership among the channel members in
order to create a distribution system that reduces inefficiencies, costs and
redundancies at the same time offers a competitive advantage, improves
quality, reliability and higher satisfaction level for customers. Supply chain
management requires cooperation throughout the entire marketing function,
including manufacturing, research, sales, advertising and shipping. In order to
achieve this, the supply chain has to coordinate and streamline the
information flow among producers, wholesalers, retailers and customers,
component-part suppliers, shipping companies, communication companies
and other organisations that participate in product distribution.
Leading and efficient organisations have been leveraging the information
sharing and flow efficiencies offered by the Internet through one or more of
the following strategies:
Enhanced Collaboration amongst Partners
In the Internet era, various supply chain partners can share information from
one another or tap into the common shared data stores. The sharing of the
customer demand data from an upstream partner can help all the downstream
suppliers in reducing the forecasting errors. Thus inventories and
manufacturing processes can be streamlined for efficiency. Further, the
sharing of production and delivery schedules of the manufacturers with
supply chain partners can lead to better material planning and thus reduction
in costs.
Material Planning
Just In Time (JIT) systems aim to minimise the quantity of inventory of
materials for the production process. In the information uncertain
environment, reliable delivery schedule requires holding large quantities of
inventory, which has cost implications, as it blocks the financial and building
capital. The benefits of JIT have derived by information technology-savvy
companies through information sharing. The Internet has made this sharing
easily accessible and achievable for all the businesses. For example, way
back in 1981, 3M Corporations discovered that 13% of materials received
were defective. So, in 1984, the Corporation decided to introduce a Just In
Time system. After the implementation of JIT, only 1% of materials were
faulty. This reduction of the faulty materials was due to the fact that the
suppliers knew that their customer, through the implementation of JIT
systems, has minimised inventory. Consequently, if a supplier provides its
client with bad materials, the client’s production process would have to stop
until new material could be found (provided the company has zero security
inventories). It is thus obvious that if the suppliers were to be blamed for the
delay in production, the client would not trust the supplier in the future.
Firms trying to implement JIT systems usually deploy a combination
mechanism like resorting to small quantities of ordering materials; ensuring
high quality or total absence of faulty materials, and frequent orders for new
materials in order to minimise the allround inventory.
JIT systems may sound simple, but they demand coordination between
supply and demand and this means that the materials have to arrive at the
plant the moment the enterprise needs them, neither earlier nor later. Thus, a
precise system that tracks the manufacturing schedule and the partners is
aware of the small orders that are about to arrive at their end, requires shared
information resources. The Internet through the innovative use of information
system automates the whole process of information sharing, manufacturing
schedule and small lot order generation at a minuscule cost per order.
Inventory Planning
The objective of inventory management is to minimise inventory costs.
Managing the inventory cost by a firm should be done in such a way that the
holding costs are minimised and so does the potential stock-out costs.
Holding costs refer to the expenses of storing products until they are
purchased or shipped to customers while the stock-out costs refer to sales lost
when items are not available or there is disruption in the manufacturing
schedule due to unavailability of stock. Of course, holding costs can be
reduced by minimising inventories, but then in case the stock-outs occur
during critical periods it may have disastrous consequences and may result in
enormous financial losses any and credibility as a reliable supplier for an
organisation. Minimisation of stock-out costs requires carrying very large
inventories, but in that case holding costs would be enormous. In order to
deal with the conflicting demands, a common solution adopted by firms is to
carry enough stock of the inventories, to cover against uncertainties in the
supply chain. Higher the degree of uncertainty, greater the stock-up inventory
as an insurance against the uncertainty. The uncertainty levels are extremely
difficult to estimate in the absence of information scarcity, thus the problem
with this approach is that it is very difficult to correctly determine the
inventory levels for each product and part.
Further, customer demands are rarely stable. In a multistage supply chain,
the variations in demand at the retail front get amplified at each stage, as it
travels upstream of the chain. This may happen due to chain partners
overreacting to the backlog orders, with little or no communication between
supply chain partners. Other common reasons for its occurrence may include
excessive time delay between order processing, demand, and receipt of
products, batching of orders to reduce the ordering costs available through
bulk discounts, reduction due to bulk transportation expenses etc., and
inaccurate demand forecasting or free return policy.
The variability of demand, caused by any of the above reasons, increases
at each stage of the supply chain, giving rise to a phenomenon called
bullwhip effect. The bullwhip effect has been observed by managers in a vast
array of industries, and in every case, it has increased both physical
distribution and market-mediation costs. The excess unplanned demand
projected due to the bullwhip effect leads to excessive costs being incurred
due to last-minute acquisition decision of the additional raw material. The
urgent acquisition of material due to falsely projected demand results in
excess inventory of unused supplies, which entail additional associated costs.
Further consequences of the bullwhip effect include, inefficient utilisation,
overtime expenses incurred during high-demand periods, further worsened by
the excess warehousing expenses incurred because of unused storage space,
as well as increases in shipping costs caused by premium rates paid for last-
minute orders.
Procter & Gamble noticed the impact of the bullwhip effect on its
Pampers diaper business, a for product which has a relatively stable
consumption pattern, as babies are consistent in their use of diapers. But the
demand at retailers, such as Wal-Mart was variable, and this increased as
orders were passed up the supply chain from Wal-Mart to P&G to P&G’s
suppliers. P&G found that the variability was self-imposed through the
supply chain’s pricing structures, incentives and planning and ordering
processes. The bullwhip effect has been experienced by not only the Fast
Moving Consumer Goods (FMCG) companies like P&G. Firms ranging from
Hewlett-Packard in the computer industry to Bristol-Myers Squibb in the
pharmaceutical industry have experienced a similar phenomenon.
The impact of the bullwhip effect and other supply chain problems can be
mitigated provided the firms are able to improve demand forecasts. This can
be accomplished through information sharing along the supply chain.
Electronic Data Interchange (EDI) has been successfully deployed to
facilitate better information exchange among the supply chain partners. The
Internet-based extranets and groupware technologies, as part of inter-
organisational information systems, provide and effective platform for
sharing the information, and thus easing the impact of the problem. EDI
involves the direct, computer-to-computer transmission of inter-company
transactions, although, in the common perception, many people think of EDI
as relating to purchasing alone. In fact, EDI involves an improvement in
information exchange mechanism for a broader set of business processes that
include credit memos, shipping documents, and other routine transactions
between companies. In essence, EDI links a company to all external parties
including suppliers, transportation carriers, public warehouses, freight
forwarders, customs clearance houses and others.
Most notable early users of information sharing mechanism to reduce the
impact of the bullwhip effect have been large manufacturers and retailers. For
example, Wal-Mart provides P&G access to daily sales information from
every store for every item P&G makes for Wal-Marts stores. By monitoring
inventory levels, P&G knows when inventories fall below the threshold for
each product at any Wal-Mart store. These data trigger an immediate
shipment. The benefit for P&G is accurate and timely demand information,
thus P&G can plan production more accurately, minimising the bullwhip
effect.
The evolution of Radio Frequency Identification Devices (RFID) to
Current miniaturised form and their cost-effective production has brought
RFID-based tracking as a new mechanism to address the supply chain
problems. RFID can improve the exchange of information between a retailer,
a manufacturer and the suppliers. Suppose that each of them uses RFID tags,
automatic alerts through the Internet, can be sent within each company and
between companies. There is no longer a need to count inventories, and
visibility of inventories is provided to all business partners that are networked
together. RFID transmits real-time information about the location of
merchandise. A retailer can use RFID to locate merchandise, control
inventory, prevent theft and expedite processing of relevant information.
Extending the Reach
The ability of the Internet to remove the physical and locational barriers can
be leveraged to extend the reach and broaden the base of supply chain
partners. Leading companies like General Electric (GE) have immensely
benefited by automating their supply chain function through forming a Trade
Process Network. The initiative has resulted in nearly 30% saving in costs
and 50% reduction in purchasing cycle time. It has broadened the scope for
GEs supply chain partners providing GE with an opportunity to interact and
reach not only immediate suppliers, but also suppliers’ suppliers and to their
customers’ customers.
Another leading manufacturer of the high tech hardware components,
Adaptec, has used the reach of the Internet to build a virtual manufacturing
facility by integrating together the processes of various manufacturers,
suppliers and suppliers suppliers on a global scale. Through the use of the
Internet technology the Adaptec coordinates in a synchronised form the
business activities/processe of all these constituents to successfully carry out
product design, product specification, manufacturing, purchase processes,
monitoring and sharing of work in progress information and shipping and
delivery status. Without the use of the Internet commerce, it would be
unthinkable to coordinate and carry out manufacturing through a globally
distributed virtual factory. The initiative has been able to save Adaptec a
roughly US $ 1.2 billion investment required to set up a manufacturing
facility. Using the virtual factory model, Adaptec has also been able to reduce
the manufacturing cycle time to 55 days, roughly half of what it used to take
using the conventional non-Internet commerce-based manufacturing
processes.
Distribution Channels and the Internet
One of the significant impacts of Internet commerce on the marketplace has
been the lowering of the interaction cost among the manufacturers,
wholesalers, distributors, dealers, retailers and consumers. The traditional
marketplace posed at times enormous barriers for the consumers interested in
the price and feature discovery of products. Also, the information flow
usually happened in pre-determined, often a hierarchical, distribution
structure. The Internet has been a great levellers by placing the manufactures,
dealers, multiple geographically distributed retailers and consumers on the
same information sharing plane, the price and product discovery barriers in
terms of incurred costs have almost disappeared.
Consequently, many novel models that innovate by leveraging on the
Internet enabled information acquisition, flow and capability of restructuring
the information sharing and flow have emerged. The restructuring of the
information sharing and flow capability has been put to use by many
companies for elimination of the layers of the supply chain. The RSNet of
Hindustan Unilever Limited was deployed to restructure the supply chain and
it successfully eliminated several layers whose sole purpose was information
aggregation, and flow in the supply chain. Internet commerce technology
made these information flow aggregators and facilitators redundant, as in the
Internet commerce environment the redistribution stockist (RS) became
capable of directly uploading and downloading the required information.
Similarly, Federal Express (Fedex) and Kinkos have gainfully created a
new document delivery system by joining their capabilities through Internet
commerce technologies. In the new model of document delivery, both the
companies are receiving the customers’ delivery documents electronically
and then these are routed to the Fedex/Kinko’s centre closest to the delivery
point. The new model bypasses the air transport fleet of Fedex, thus
achieveing savings in cost and delivery of documents on the same day.
The restructuring of the supply chain often results in shortening it and, as
a consequence the firm is able to improve on order-to-delivery time,
reliability, broaden the product choices, reduction in the costs and better
profits for the firm. When showered with these advantages, the customers
respond, resulting in increased price realisation and market share for the firm.
The information aggregation capability offered by the Internet can also
help in aggregating the supply with demand. Many new players with
negligible physical infrastructure become the aggregators of either the
demand or supply carry out the task of information mediation. These range
from pure infomediaries that have negligible physical infrastructure, to hybrid
intermediaries who rely on both infomediation and some elements of physical
distribution.
Ebay (http://www.ebay.com), Auction India
(http://www.auctionindia.com) have been pioneers in the field of information
mediation and are a great example of the first category. These players have
created virtual marketplaces that primarily aggregate the demand from
scattered buyers and the product listings from scattered sellers without even
handling or holding any merchandise. In these model customers, the
shipment is carried out through logistics intermediaries, who, in turn, derive
scale economies based on the traffic in their networks.
The second category, hybrid intermediaries, include most of the online
retailers like Amazon.com (www.amazon.com), FabMall
(www.FabMall.com), JC Penny (www.jcpenny.com), Office Max Online
(www.officemax.com). These players do leverage on the information,
mediation aspect, but also back it up with physical infrastructure through a
warehouse and sometimes even a store network. For the growth of these
hybrid infomediaries the existence of third party logistics intermediaries
becomes imperative. The logistics intermediaries work on economies of scale
by aggregating the logistics services for several such firms and thus are
capable of offering superior ability for moving shipments around the globe.
Finally, some argue the Internet technology has brought the “end of
distance” and the homogenisation of time in modern retailing. Indeed, before
the arrival of the Internet, geographical isolation was one of the major
reasons why international commerce could not easily develop. However,
nowadays, thanks to the Internet technology in product distribution-
especially those that can be digitalised, such as pictures, videos, sounds and
words, distance has no longer any effect on costs. The same is true for
services.
PRICE
The availability and reach of Internet marketing has resulted in extreme price
competition for goods and services that are perceived as commodities, due to
factors that might permit price premiums such as store location, availability,
are absent and also because of the relative ease of comparing prices at
different websites. For example, e-campus offers cheaper textbooks than
bookstores at many campuses. The agent-driven Shopbots, coupled with the
information push economically viable on Internet channel, is likely to
accelerate the elimination of price differentials among goods and services, as
the push technology allows customers and further markets to:
Subscribe to channels which monitor the price changes of competitors;
and
Disseminate their competitive pricing responses instantly to consumers.
Also, in several Internet business models, specially the ones that are based
on digital or information goods/service, the value creation in the chain, unlike
the manufacturing situations, is not vertical at all. In a vertical value chain,
the activities follow a hierarchical sequence, where each element of the
hierarchy plays a vital role in terms of facilitation of material and/or
information flow in both the upstream and downstream directions. As
discussed earlier, electronic commerce alters and greatly enhances the
information sharing and flow capabilities and substantially lowers the cost
among all the players in the value chain. The electronic commerce market
space emerges as a platform where any of the constituent elements of the
value chain can be arranged in various linear and non-linear structures,
including the possible elimination of a few of the elements. Thus, the Internet
driven commerce gives an opportunity to firms for streamlining their
coordination and distribution costs. The price of a product in a market place
consists of the following the elements:
1. Production Costs
2. Coordination Costs
3. Profits
4. In addition to these elements, customers also incur the search cost
The streamlining and business process restructuring initiated by the
assimilation of information technology can assist in lowering the overall
production costs of a firm. Much of the impact and improvement in
production costs are due to information integration and efficient
dissemination features offered to enterprise integration, enterprise resource
planning applications resulting in the adoption of the best practices.
As described earlier, Internet commerce has a significant impact on
streamlining and restructuring of the supply and distribution chains, making
them more efficient and thus reducing the overall coordination costs.
Moreover, e-commerce consists of selling goods directly to the customer
without passing through retailers and distributors, hence cutting, costs by
avoiding intermediary margins as in offline shopping. By consequence,
online prices are even lower as a company deals directly with its clients. This
can be observed for example on “la fnac” web site–www.fnac.com–where
online books cost less than offline books. This due to the fact that e-
commerce also does not require a direct sales force Another important point
to mention is that the ordering process has been changed, as the invoice
processing costs less; other expenditures linked to catalogue editing and
printing can, as well, be saved.
It is worth noting that some online companies are not physical companies,
hence avoiding the fixed expenses linked to this kind of structure and
enabling them to sell at fairly low prices. I-tunes’ success has been built on
this peculiar aspect, selling digital music while charging a minimum fee: a
song in i-tunes costing 0.99$ and an album only 9.99$ which is by far
cheaper than actually shopping offline for CDs and other multimedia
features. It is, however, important to mention that the arrival of the internet
created illegal programme song downloads as people can nowadays
download, illegally or not, software for free from different websites or by
peer to peer. To fight this phenomenon, companies have been forced to lower
their prices so that people would buy their product even though they can
download it for free. We have seen through the year that a lot of programme
have lowered their prices.
The e-commerce also significantly alters the profit components, due to
free availability of competitive product comparison and information. The
information availability and global reach has made the role of ecommerce in
determining the product prices quite significant. With the growth of the
customer segment using ecommerce, one of the prominent elements of the
four Ps, i.e., price has severely come under pressure as marketers have to try
to set the prices that will match the globally informed buyers expectations
and ensure that customers clearly either see a price or value added feature
differentiation and yet operate above the floor under which no profits are
made.
With the easy availability of price and product feature comparison
information in the e-commerce environment through a price comparison site,
the customer has become an informed buyer. Today, a customer can refer to
price comparison websites where offers of different retailers and for the same
product are grouped and evaluated, making it easy for them to match the best
deal. Not only does this comparison entail evaluations for similar products,
but it can also be done between competing products, thus permitting buyers
to see the difference between the different characteristics of the goods and
services, their strengths and weaknesses, as well as their prices. An example
of such websites is PriceGrabber (http://www.pricegrabber.com), a
“Comparison Shopping beyond Compare”.
Therefore, companies should be wise when pricing their products because
customers have gained in awareness through the emergence of new tools on
the Internet. If a product is too expensive compared to its competitor, the
customer will simply opt for the cheapest product!
ILLUSTRATION 12. 4 PriceGrabber.com
PROMOTION
If we take note of the overall influence of the Internet commerce on the
marketing domain, it does clearly appear that the promotion of goods and
services has been largely influenced by the Internet as a medium of
communication, as well as a commercialisation tool.
Online marketing, and the different processes that are related to it, has
undoubtedly generated a reshaping of the way the promotion is made to
businesses and consumers: from the simple e-mail advertising all the way
through e-marketing strategies, the effects of e-marketing have been giving a
new dimension to the use of technology in conducting customer relationships.
The Internet with a growing numbers of subscribers has emerged as a
powerful platform for reaching large audiences and the delivery of brand
messages. Through the two-way communication and interaction ability of the
Internet, it is also worth noting, that it also allows consumers to widely
spread critical opinions and experiences all through the web-making it an
extremely sensitive platform.
A quick glance at the trends in Internet advertising in the United States
shows that it has garnered a far wider acceptance, from the first banner ads on
hotwired.com in 1994, to the extent of crossing the global expenditure of US
$ 21 billion in 2007; a noticeable expansion that is to go with the idea of
restructuring the company’s marketing expenses.
Electronic marketplaces also offer endless opportunities to promote a
company and its products or services. With its ever-growing pool of middle
to upper class users, the Internet provides access to prime target groups. In
addition, at a fraction of the cost of traditional means such as print, television,
or radio, online promotion can be delivered almost instantaneously around
the globe. Several studies, including the ones conducted by Internet
Advertising Bureau (http://www.iab.net), IBM indicated that firms putting
forth online catalogues on the Internet could save up to 25% in processing
costs and also reduce the cycle time by up to 62%. Therefore, its not
surprising to see that Internet’s fastest growth has been witnessed by
advertisement and marketing.
The Internet emergence as an interactive platform for promotion of goods
has greatly influenced both the digital as well as tangible products and
services. In the case of digital products and services available through
electronic commerce businesses, the traditional mass-market approach of
creating hype and brand-building through television, billboards, and print
media promotion has limited role, as the audience utilising the electronic
commerce channel is far better educated and aware than the average
consumers. The audience engaged in e-commerce transactions for digital
products and services requires a greater deal of interactive information prior
to making up their minds regarding the utility, the traditional approaches with
the limitation of one-way communication can assist in creating awareness but
are grossly unsuitable for promotion purpose where interaction is almost a
necessity. Also, the two way communication channel provided by Internet is
not only useful for digital products and services, but also empowers the
customers in eliciting the appropriate information for the tangible goods and
service.
The traditional mass market approaches are push-driven. In contrast, the
Internet medium is capable of both the push and pull driven approaches.
Further, on the Internet, even in the push model (practised through e-mails),
the consumer still has control on the type, duration and exposure. In other
words, on the Internet the consumer has a choice to visit a message and spend
time in exploration of the message, depending on his desire how much time
to spend on it. Marketers need be aware of the transformed communication
perspective in the Internet-enabled commerce environment, in order to utilise
it for improving the effectiveness of promotional campaigns. The Internet
commerce environment impacts on the following promotional strategies and
mixes.
1. Advertising
Since 1994, when the first banner advertisement appeared online, the online
advertisement industry has grown to US $ 21 billions through a series of
innovations. The online banner still remains a measurable and effective
means both in terms of costs and recall. The traditional marketers plan and
buyout the campaigns on the standard media of TV, Radio, print, billboards
etc., where the costs are determined by rate cards. In online environment, the
media consists of all the cyberspace which can be targeted by emails, or
places that are able to aggregate the cyberspace visitor. These websites or
Internet forums commonly include portals like MSN.com, community
websites like myspace.com, search engines like Google through sponsored
links, shopping agent oriented websites like pricescan.com, blogs, message
boards and chat rooms. The common pricing models for online advertising
include Cost per Thousand Impressions (CPM), Cost-per-Click (CPC) and
Affiliate revenue also called Cost-per-Action (CPA).
Further on the Internet, firms can create interactive rich advertisements to
appeal to customers by providing them with the exact information they were
looking for. In the interactive advertisement, model consumers can simply
wade through the web information by clicking on icons or hypertext to gather
the information. Consumers may select to go through the detailed product
information in the form of text, picture, audio, or video at their pace. The
approach is a highly effective way to reach consumers who generally do not
like the mass-market hard-sell approach. The non-intrusive and user
controlled (pull-based) advertisement is likely to work best for the informed
consumers.
2. Sales Promotion
The objective of sales promotion is to facilitate the movement of product
from producer to consumer through short term incentives. The Internet, being
a two-way dynamic channel, can be used by marketers in designing effective
sales promotions in the following ways.
First, Internet commerce being an interactive and dynamic environment,
enables the marketer in designing more innovative and sticky promotion
schemes. These schemes involve lower costs and do not clutter up the
physical mail boxes of individual customers. The markers can use the
creative aspects in designing rich media-based promotions that are not only
informative but also enjoyable to consumers. Thus, giving the control in the
hands of consumers to download, play or interact with only what is of interest
to them at time slots that suits them.
Secondly, Internet commerce being highly driven by the database/
information servers at the back end, also provides the marketers with a great
opportunity to profile the consumers and offer a high degree of personalised
promotion.
3. Personal Selling
One of the most commonly used techniques for recommendation generation
is collaborative filtering. Collaborative filtering identifies a subset of users
that have similar tastes and preferences to that of the target user to generate
recommendations. More specifically, collaborative filtering process involves
three stages viz. (1) computing similarity between the target user and all other
users, (2) selecting a subset of collaborative users based on the similarity
coefficients computed in step 1, and (3) offering recommendations based on
products liked by collaborative users.
Data mining is another technique used for recommendation generation.
Data mining is defined as a non-trivial process of extracting potentially
useful, interesting, and actionable information from massive databases.
Specific data mining techniques used for recommendation generation include
association rule mining, clustering, web mining or a combination of them.
Information retrieval is yet another method for recommendation
generation. There are variety of shopping assistants available on the web
(such as Bargain Finder, www.bargainfinder.com; Dealtime,
www.dealtime.com; Shopping, www.shopping.com; E-pinions,
www.epinions.com) that use information retrieval-based methods. These
shopping assistants provide an agent-based shopping support for customers.
They take price and a set of product features as inputs, and match them with
available products on the Internet to select a set of products of interest to the
customer. These agents also provide services such as product ratings,
customer reviews, price comparisons and details of product availability
across stores. However, selecting suitable products in the vast Internet is a
challenging problem. Other shopping agents available on the web such as
Active Buyers’ Guide (www.activebuyersguide.com) take into account the
importance of product features in addition to the feature itself to select
products of interest to the customers. In essence, the shopping assistants or
agents available on the web use a set of customer-desired features, and match
the same with the available products on the web to select a set of products for
recommendations. The recommendations generated in such systems are
generally product variants rather than cross-category products as in
collaborative filtering-based methods.
The coupling of Internet commerce with database information offers
unprecedented opportunities to create a promotion mix that caters to
individual requirements and thus fosters a long-term relationship with the
customers. Unlike other mediums, the Internet fosters conversation and thus
in the Internet commerce era, companies must be ready and willing to listen
to consumers and engage them in conversation. In the first phase of Internet
commerce, the companies fostered this by strategically locating the e-mail
button or feedback boxes to elicit the comments and views of the customers
regarding their products and services. The dedicated product-user groups,
blog sites are also some of the means utilised by firms to foster the
conversation. These company-supported sites are often the places where
consumers openly discuss the product failures, flaws, fixes and work-around.
The firm’s customer support personnel participate in these forums to offer
technical advice, explain the reasons and future directions, or to simply calm
the customers while a solution is being work out.
MARKETING COMMUNICATION
While considering the increasing number of ways of dealing with e-
marketing, it is important to look at the line of action that a firm may deploy
to elaborate an online marketing strategy.
Electronic marketing is more than building a website or promoting a website
as at the backend of the website is a real organisation with real goals.
Thus, the Internet marketing strategy charted out for any organisation must
include various aspects of online advertising products, services, and websites,
including market research, email marketing, and direct sales. Depending
upon the business model and its stated goals, the Internet marketing strategy
has be designed and appropriately aligned with those goals. However, the
reach and access of the Internet and its ability to amplify the message has to
be managed with extreme care. Marketing communication and public
relations become extremely important, as the messages on the electronic
markets travel wider and faster.
The Integrated Marketing Communication (IMC)
It has to be viewed as a cross-functional process for planning, executing, and
monitoring brand communications designed to profitably acquire, retain and
grow customers. By specifically studying the variety of offline and online
media advertising, integrated marketing communication allows a firm to
produce the best marketing mix between these two types of communication.
Successful firms can thus, by using technology, pay more attention to high-
value customers and develop high analysis techniques linked to databases
formed through studying of their consumers online behaviour.
Marketing Public Relations
Marketing Public Relations (MPR) requires a methodical planning of
activities to ensure that the key messages that serve the business goals are
effectively communicated to target audiences. The important aspect is to
identify, what your key messages in consonance with business goals are, and
what is the most effective way for communicating them to target audiences in
order to bring about the desired attitudinal change. These aspects acquire an
added significance in the internet environment that has capability to multiply
and amplify the message manifold instantly, thus a slightest deviation may
also get multiplied and can have disastrous consequences.
The MPR concept that includes the building of awareness about a brand is
an important tool for capturing the influence of public opinion. MPR in
internet commerce is directed towards the firm’s customers and prospects
mainly through the use of a website—serving as an electronic brochure. On
the electronic markets thus improving the customers online experience
therefore the becomes the highest priority. It is in this perspective that firms
usually use a web content that includes press releases and publicity.
There are several advantages of using these methods through the web for
publishing products or services information: first, the internet is a low-cost
alternative; second, the information is often updated; third, the impact of
update is reflected instantly; finally, the Internet always attracts new potential
customers who are searching for a particular product. Hence, the online
marketers have to continuously pay attention to the Search Engine
Optimisation (SEO) so as to allow the information on their products show up
in an optimal manner. In the search business, it is the top 10 results on a
search result page usually capture more than 78% of the traffic. Obviously,
firms should identify and then carefully emphasize elements that would lead
their pages and advertisements to appear in the top 10 results for the target
audience—thus delivering a competitive e-marketing advantage. The
measurement is an important aspect and the Internet provides an easy way to
track the traffic. The marketers should develop a plan that addresses the roll-
out of a campaign with a tracking mechanism to ensure the impact of the
programme. The web analytics provide excellent ability to measure the
impact of the programmes.
Although resulting in many benefits, e-marketing, as any process, suffers
from some limitations that range from technical issues such as heavily loaded
web pages, slow Internet connection to other psychological (disturbing
influence) or law-enforced issues.
In essence, the effect of online communication should be seen as a
juxtaposition of the basic marketing concepts with the field of digital
technological advances. For online marketers, the Internet has emerged as an
inevitable element of growth. The online marketers can take the usual needs
of the customer and transpose them in accordance with their business goals to
harness the various advantages available on the Net and the growing online
buying practices.
Thus, online marketing has not only brought up new methods that parallel
the established practices, but, has also led to the implementation of new
practices in the existing marketing domain: first, by creating an inherently
profitable technique to catch up with the mass market; secondly, by making
of the Internet space a business (electronic marketplace for transactions) itself
as, many of the leading e-commerce websites are generating large sums of
revenue through promotional links and advertisements.
SUMMARY
Across the different elements of the marketing mix, we can see several areas
where the Internet’s strengths stand out. These include possible cost savings
through digitalisation of communication, as well as reduced transaction and
search costs. Further, the ability to talk to customers—both businesses and
consumers—in a more interactive and customised manner enables firms to
add more value to transactions by changing the marketed offer. The Internet
continues to provide a wide range of opportunities for firms to step in and
offer new products, new service-augmented products, innovative promotion
techniques and pricing strategies. Given the expanding nature of the Internet
(in terms of scale and scope), there are still numerous untapped markets for
products as well as access to new markets.
Perhaps the lesson that can be learned from this analysis is that although
the Internet has made a definite mark on marketing in all sectors, traditional
business is not dead and its business principles still apply. The challenge of
marketing managers in the future will be to recognise how to use the Internet,
along with the existing processes, when appropriate, to move strategically
within and between markets.
This is not to say that the Internet is not without its corresponding
weaknesses. While its digital nature enables new distribution methods,
traditional logistics and supply chain management problems cannot be cast
aside. Further, the consistency of all the four Ps in the offline and online
world means that different strategies are needed in terms of branding and
positioning as the competitive environment has also shifted because of an
increasingly free flow of information. Threats to previously stable businesses
have emerged, thanks to the ability of consumers to exchange price and
quality information on the internet and compare, often in real time, different
available offers. The transnational nature of the Internet further amplifies
these challenges and appears as an open door to businesses growth prospects.
REVIEW QUESTIONS
1. What are the main challenges faced by marketers in the Internet age?
2. How has the Internet technology influenced the supply chain
management?
3. What impact has the Internet had on new products?
4. Why do you think distribution acquires greater importance in electronic
Markets?
5. Why do you thinks the Internet and rich media has accelerated the
growth of viral marketing?
INTRODUCTION
The flourishing electronic commerce environment requires technological
solutions and support that provides a secure and interconnected cyberspace.
The vast number of people, connected on cyberspace, present a great
marketplace for merchandisers, traders, and manufacturers. The world wide
web technology has already proved its viability for information publishing,
multimedia content creation, and distribution over the cyberspace.
Consequently, a plethora of businesses have popped up to service the needs
and requirements of people connected over the internet. In this exploding
world of cyberspace, trying to locate the information, the service provider, or
merchandiser, that may meet the requirements of a client, is an extremely
challenging task.
The traditional mechanism of surfing the cyberspace for locating
interesting information sources does not scale well. In the surfing approach,
one types the name of a site, browses, and on finding interesting links on the
web page, clicks on the same. In the approach information seekers follow
links taking them from page to page making occasional educated guesses
along the way. Surfing works well, when the size of cyberspace is limited or
the information seeker has no paucity of time. In the real world, the
cyberspace has already grown to millions of sites and people like to locate the
items of interest quickly yet accurately, therefore Search and location
services have become the key enabler of the electronic commerce.
Search engines and directory services have emerged as two popular
mechanisms, that fulfill the need of basic business infrastructure, required for
locating a business/service/information of interest to users. Directory services
provide a mechanism wherein web sites are organized based on subjects.
Search engines typically index all web pages based upon the content of web
pages, thus, offering the ability to find relevant web page addresses, based
upon keywords.
INFORMATION DIRECTORIES
Directories on the internet carefully organize the internet resources in a
hierarchical structure, that lends itself to browsing. The directories offer
services that are typically offered by business directories of telephone
companies such as Yellow Pages. In the simplest form, directory service lists
the web sites in alphabetical order and links it with the URL of the web site.
The hierarchical directory structure consists of several levels, starting from
top level classification, sub classification within each class, and further sub
classifications.
Directory Organization
Popular directories organize the information in a hierarchy of categories and
sub categories. Each sub category may have other subcategories and/or the
links to web pages that are the best source of information on a given topic.
One of the most common directory organization methods is based on
subjectwise classification. It provides a structured and organized hierarchy of
categories for browsing the information by subject. Each category and/or sub
category offers links to other sub categories, and appropriate web pages
(URLs). The directory administrator, through the input from
editors/reviewers, assigns categories to web pages. Depending upon the need,
the administrator may divide/sub divide various categories and create newer
categories. Many subject based directories also support keyword searchable
indices, at each level to assist the user in locating information within a rather
large directory sub tree. These indices are not based on the full text, but
simply on the information you see in the directory, i.e., titles, brief annotated
description, subject category and so on. In other words, it consists of what
you see on screen in a directory listing. There is no standardisation of
subjects, they vary with the intent of each directory service, and can be
created by the directory administrator as and when found suitable. The
subject categories, sub categories, and web page entries that are included
under a sub category are through the human selection and review process. In
other words these directories are built manually. As such these directories can
vary from a small to a large size, depending on the scope, but tend to be
smaller than full text based search engines.
The subject directories, depending on the scope, come in variety of
categories—general purpose directories, academic directories, commercial
directories, industry-specific directories, and portals. Yahoo! one of the most
popular directories, organizes information based on subject trees and offers
links to web pages, with brief annotations. WWW Virtual Library is another
subject tree directory service, maintained through its volunteer’s efforts.
These directories are large, with millions of catalogues pages, and place
minimal restrictions on the material accepted for inclusion. Another set of
subject based directories, such as GNN’s Whole Internet Catalogue,
Magellan, also known as McKinley’s Internet Directory, and Point
Communications, provide significant value additions to each link with
commentaries and ratings provided by skilled reviewers.
The operational architecture of a directory is shown in Fig. 13.1. The
directory server manages two important databases. The first database
organizes and stores the subject tree structure in a hierarchical form. The
second database maintains a searchable index of the title, subject, and
annotated information available in the subject tree. A web site that wants to
register with a directory contacts the directory server to submit the required
information for including its URL. Typically the required information
includes a brief description, keywords that describe the site, the URL, and
useful categories and sub categories that user believes are suitable for the site.
A reviewer/editor visits the site for evaluating and verifying/identifying the
appropriate description, categories, sub categories etc. of the site. The site is
added to the subject tree using the information provided by the reviewer; the
information is also used for updating the searchable index. From the user’s
perspective, the information can be searched or located using the searchable
index or by browsing the subject tree. A directory server offers a browser-
friendly interface to its clients, for browsing the subject tree. Users move
from category to sub categories by clicking on the appropriate sub category.
Interested users can also find links to relevant websites by typing keywords
as the search input. The input is used for searching the searchable index
database, and returns links to relevant web sites.
SEARCH ENGINES
Search engines are massive databases that store inverted indices of text words
and web page addresses. These index databases are assembled through an
automated mechanism. Search engines gather information about web sites
and organize them into efficient, searchable structures. The enormous size of
the information available in world wide web requires tremendous computing
power and organization of information in efficient searchable structures, in
order to service the queries of clients’ in a reasonable time frame. Some
search engines handle this issue by curtailing the amount of information they
absorb from a web site. Although this approach makes data organization and
computing power requirements more manageable, it may lead to search
results that definitely include web sites that contain included relevant
information, but exclude sites that may have relevant information, but in the
part that was ignored during the information collection by the web server.
Search Engine Classification
Search engines can be broadly classified into two categories. The first
category of search engines can be characterized as those which collect the
information from the WWW on their own, and organize, store, and manage
their indices. Since, these engines compile and manage own databases, they
require larger storage capacity and computing power. These search engines
are called primary search engines. The second category of engines do not
their compile own databases. Instead, they operate on databases of multiple
search engines of the first category. These engines search multiple databases
of the primary search engines, simultaneously, and then rank the results by
combining the multiple streams. The second category of search engines are
referred to as metasearch engines.
Search engines offer users the facility of finding relevant web sites, based
on keywords, phrases, quotes, and information buried in the full-text of web
pages. Since, full-text search engines index almost every word, they are
capable of retrieving tons of documents that may have some relevance to the
topic being searched. These search engines are capable of returning a wide
range of responses to specific queries.
The scope of search engine queries covers a very large portion of the
publicly available pages on the exponentially growing Web. The difficulty of
categorizing the enormous number of resources available on the web poses an
extreme challenge to manually reviewed directory structures. Search engines
are the best available mechanism devised for finding and locating information
on the web. The enormity of this problem has an overbearing impact on the
scaling up of the traditional library cataloging mechanism. As of now, the
search engines are the only mechanism that can index all the information in
internal structures, and use techniques that have evolved over the past couple
of decades, in ranking the relevance of information in text based databases.
On the other hand, the large number of web resources and almost all the
words in each web resource, indexed by search engines, increases the
probability of hundreds of thousands of irrelevant responses. Since, all the
words are being indexed, it is highly likely that queries will return lengthy
documents in which the keyword or phrase appears only once.
Each search engine offers clients a user interface, loaded with various
search options. The user interface provides users with an ability to express
the query in terms of key words, phrases, boolean expressions and, in some
engines, the scope of the search as well. The back end uses software
programs to search indices for matching keywords and phrases, and the
findings are presented to the user, in a ranked order. Various techniques that
have been developed in the information retrieval area are applied for
determining the relevance ranking of a document. Although the objective of
the software programs may be similar, each search engine differs in terms of
size, speed, content, and relevance ranking schemes. Therefore, the search
experience will be different on every engine the client uses. The difference
may not be a lot, but it could be significant. Recent estimates put search
engine overlap at approximately 60 per cent and unique content at around 40
per cent. The other important aspect of search engines is the mechanism they
use for gathering information, for building the index databases.
Information Collection in Search Engines
Search engines gather information from web sites and create a database to
store it as a text index. Search engines employ programs that automatically
crawl through the cyberspace, visiting site after site. These programs are
referred to as the “crawlers”, “spiders”, or “robots” (“bots”). The first
crawler, called World Wide Worm, was created in 1993. It crawled through
the cyberspace by visiting one site, gathering and indexing all the information
pages and then hopped to the next site by following a link in the existing site.
Crawlers, spiders or robot programs traverse through cyberspace from link to
link, identifying and perusing pages. In the process, sites that do not have
links from other pages may be missed by spider programs. Once the spiders
get to a web site, they typically index most of the words on the publicly
available pages at the site, creating a huge text index database. Eventually, if
the cyber space is fully inter linked, all the pages on the cyberspace become
part of the text index of the search engine. Many pages, especially newly
created web pages, may not have links from existing pages that are part of the
search engines text index; in such cases, web page adminis- trators can
submit their URLs to search engines for “crawling”, and eventual inclusion in
the index databases.
Fig. 13.2 Internet Search Engines
The crawlers/robots crawl through the internet, continuously, in order to
index as much of the latest information as possible. Typically, robots revisit
indexed links on a periodic basis to keep the information in text index
databases up-to-date. During a revisit, a robot may find dead links, which are
then removed from the index databases, or additions and changes in the
information, which are duly reflected.
Robots have been used extensively by search engines. In addition to
search engines, they have also been used for the purpose of creating mirror
sites and keeping the content of these site up to date. Software archives and
bibliographic databases are typically mirrored in several sites, across the
continents, to reduce the load on a single site, and provide acceptable levels
of performance. Robots are crawlers that automate the task of information
collection, and are useful in gathering information and maintaining index
databases. Once a crawler visits a web site, it will collect all the information
available on the web site, for indexing purposes; much of it may not be very
relevant. Robots lack the intelligence to analyze the information, therefore,
they may also add information that is not so relevant to the index database.
For example, a robot visiting a web page with links to common gateway
interface (CGI) programs stored in/cgi-bin/directory may collect the
programs and make them part of index database. In order to address the
above problem, a standard, called Robots Exclusion Standard, has been
defined to specify robot behavior when crawling through cyberspace. Most of
the internet robots have adopted the exclusion standard. As per this standard,
in case a web site administrator wanted to exclude certain documents from
inclusion or even all the documents on the site from inclusion, in the index
database of the visiting robot, this can be specified in the robots.txt file. The
visiting robots look for the robots.txt file in the root directory of the web
server. For example, in the site http://www.yoursite.com, the robot will read
the content of document http://www.yoursite.com/robots.txt and act
according to the content of the document. The robots.txt follows a very
simple exclusion standard protocol. The exclusion/inclusion is specified
using two directives “user-agent” and “disallow”.
In order to exclude the content of cgi-bin/ directory from all the robots, the
robots.txt file can be constructed as follows:
User-agent: *
Disallow: /cgi-bin/
In the above example, the value of “*” for user-agent implies that it
applies to all robots, while the disallow field specifies that the content for
directory tree cgi-bin is excluded. It is important to note that regular
expressions are not interpreted in either of the two fields-user-agent and
disallow. The value “*” in the user-agent field is a special symbol implying
that it applies to all robots. But, the file cannot contain lines such as
‘Disallow: mydocs/*’ or ‘Disallow: *.gif’. The following content will
exclude the web site from all the robots.
User-agent: *
Disallow: /
The following file will permit access to all the contents by all robots. Even
the absence of the robots.txt file on the site will have the same effect.
User-agent: *
Disallow:
The following file will exclude selective documents and directory contents
from all the robots. Each resource exclusion has to be specified in a separate
disallow line.
User-agent: *
Disallow:/cgi-bin/
Disallow:/~bhasker/index.htm
Disallow:/private
The following example excludes a single robot named in the user-agent
field (badcrawler) from accessing all the resources (specified by /).
User-agent:badcrawler
Disallow:/
Or the robots.txt can specify access to the resources, only to a single robot
by specifying it by name, as shown in the following example.
User-agent:WebCrawler
Disallow:
Thus, whenever a web search is performed through a search engine, it is
done using the index of sites that match the clients’ keywords and phrases,
with those in the texts of documents that have been visited and indexed by
the engine’s database. In other words, the search is limited to the part of the
web space that has been visited by the search engine sometime in the past,
and not the entire cyberspace in its current and updated form, as. Robots try
to keep the databases up to date, but in some cases the information may be a
few weeks old, or even older.
Major Search Engines
Searching and locating of the relevant information on the Network has been
an important issue even prior to the arrival of World Wide Web. The Archie
developed by students of McGill University was one of early attempts of
using a software tool for locating the information on Internet. It was later on
followed by several other tools like Gopher, Veronica and Jughead. With the
growth of World Wide Web pages on the Internet, in 1994 a crawler based
search engine called Webcrawler was launched. This was soon followed by
several other crawler based search engines, namely, Excite, Lycos and the
first human powered directory, Yahoo!. With the growth of web, a need for
better technology for indexing and searching became imperative a next wave
of search engines like Infoseek, Altavista, HotBot, Ask Jeeves and more
recently Google were launched to meet the requirements. According to
Nielsen/NetRatings studies as of 2007, much of search queries are directed
towards few search engines. The Google ranks on the top of the list of most
popular search engines with a whopping share of 49.2% . The other four in
the list are Yahoo with 23.8% , MSN with 9.6% , AOL with 6.3% and Ask
with 2.6% .
SEARCH ENGINE MARKETING
Search engines and directory services are the most popular methods of
locating the relevant information in the cyberspace. Most of the web surfers
primarily utilise the search and directory services web sites as starting point
of the surfing. The enormity of information available on the web poses a
serious challenge for the information seeker. The search engines and
directory services are the mechanisms that help in addressing the challenge
by indexing and/or categorising the available web pages on the Internet. With
continuing growth, the number web pages on the Internet have already
reached the order of hundreds of billion. In such a mammoth pool of
information, locating a set of relevant information based on the user’s query
consisting of few keywords is an ominous task. A simple search for query
“mp3 player” in Yahoo, results in a set of 102 million web page references.
Thus, need for a service that can filter out the irrelevant pages from result set
and mark them on authenticity, reputation and relevance becomes imperative.
Thus, a user can have a look at references to 20–30 web pages on the result
page that are authentic and relevant. Having emerged as a necessity, the
search services have become major internet traffic aggregators. With Google
alone receiving and answering roughly 10 billion queries in a year, people
cannot afford to overlook the vast marketing potential of search engines. In
order to harness this potential, the marketing managers must be in a position
to understand and appreciate the ways in which search engines index and
retrieve the data collected from the web pages. A better understanding of the
search engine operation can help marketers in tuning the content and design
of their website in such a fashion that their sites get ranked in top 20 results
for the right set of query words. The objective of achieving the top 20 rank
for relevant queries and keywords requires a better understanding of the
techniques used by various search engines such as Google, Yahoo, MSN, and
AOL.
Initially the search engines started off as a pure indexing service that
included the contents of pages from various sites and generated revenues
through banner advertisements. But, over the past decade, as the website
content creators became more sophisticated and exploited the indexing
techniques used by search engines to attain higher ranks, the search engines
have also evolved to cater to search engine marketing through innovative
revenue generation models. The search engines have also become more
sophisticated in scaling up to match the growth of web content by innovating
newer indexing techniques, criteria for what portion of content becomes part
of the index mainly due to the following:
1. The excessive number of websites and the amount of information
available on these websites has brought the quality issue to the fore. As
a consequence, the search engines have to figure out not only the
websites that match the queries and keywords but also determine the
authoritativeness of the information present in the website during the
ranking before including them in the top 20 ranks.
2. The search engines have to be able to detect the practices adopted by
search optimisers to achieve higher page ranks, using dubious means
called search engine spamming.
The visibility of a web page in Search Engine Result Pages (SERPs) has
provided opportunity to a new group of intermediaries, called Search Engine
Optimisers, who can promote your website to attain higher ranks for
appropriate guenes. The search engines have also grown sophisticated in
revenue generation models. Gone are the days when all the content indexed
by the search engines and directory services was for free. In the following
sections, we will discuss several newer revenue generation and sharing
models that have appeared over time and also various search engine
optimisation techniques that need to be understood and are utilised by the
search engine optimisation community to get better placement of web pages.
Revenue Models
As stated earlier, the initial crawler based search engines scanned the content
by crawling through the World Wide Web and created the index that was
searched by users by typing their queries in a search window offered by the
homepage of the search engine. The search engines generated revenue by
placing banner advertisements.
In 1998, GoTo, later branded as Overture, launched a Pay Per Click (PPC)
model. In this model organisations were provided with the ability to buy their
ranks through the process of bidding for the click-throughs. Since then,
Google, Looksmart all joined the fray and thus the results from a search
engine began to consist of two categories, namely, ‘Paid Search’ and ‘Pure
Search’. It is quite common today to see the paid advertisements appearing at
the top, right after the query, followed by primary query results. Thus, the
placement of result web pages has become dependent on the marketing
budget spent on it. In order to promote the web business through search
engines, it is essential to understand the common fee models that are in used
for promotion of web pages. The common ones include:
(a) Pay-Per-Click
(b) Pay For Consideration
(c) Pay For Inclusion
These revenue models are explained further in the following subsections.
Pay Per Click
In this model, the search engine sites are willing to list your web site at the
top of the Search Engine Result Pages (SERPs) for a price. The model allows
the web marketing managers to select the most popular keywords or phrases
against which they would like their site to appear on the top of result pages.
The model, also referred to as sponsored search, has been put to use by major
search engines such as Yahoo! Search Marketing (also known as overture),
Google Adwords, Looksmart and MSN.
In this model the website managers identify the popular keywords and
phrases against whom they would like the advertisement (listing) to appear
and then place a bid for the keywords, like in an auction. The website
managers sign up for these search marketing programmes of Google, Yahoo!
and others as the case may be and deposit the money in the account.
Although, various search engine marketing programmes differ on the ways,
they manage and charge for the bids the general procedure is as follows:
1. E-Marketer (website manager) signs up with one the search engine
marketing service providers and deposits the money in the account.
2. The website manager selects the keywords and phrases, which are
highly relevant to the website content and are likely to yield right kind
of customers, and bids for them in an auction style.
3. Whenever a surfer types one of these keywords or the phrase for which
the website manager has put in a bid, presuming it to be high enough for
the top spot, the search engine displays the website advertisement in the
sponsored/paid results area usually right below the query window at the
very top of all the results.
4. In case the surfer/searcher decides to click on the sponsored link and
lands at their website, the bid money is then deducted from the deposit
account of the website. The deducted bid amounts vary widely from US
$ 0.10 per click to US $2 per click for some hot keywords. However, no
money is charged for appearing in the paid listing area.
5. In case a competitor raises the bid for the same keyword, then the
previous company, gets downgraded and the competitor’s website
appears in the paid listing area. The downgraded company is informed
that they have been outbid. Consequently, they may decide to raise their
bid further. Sometimes, it may lead to price war among the competitors
for some keywords, leading to non-tenable return on investment (ROI)
for the companies involved.
Most of the companies use an automated bid management tool for managing
the competition and optimise in the event of a bidding war among
competitors for the same keywords.
Pay For Consideration
In the case of humanpowered directory services that review the content of a
website prior to categorising them and including them under the directory
path, in some cases it may take several months for the human editors to
access and decide on whether to include a particular website or otherwise.
The Yahoo! and Looksmart directory services have introduced a payment-
based service that ensures the site will be accessed, reviewed and the decision
will be made in a specified period of time. The ‘Pay For Consideration’ only
assures the company, which is submitting the URL along with the payment,
will be considered and reviewed in a specified amount of time. However, no
refund of the amount charged is made irrespective of whether the decision is
made to include or not to include the website in the directory index.
For example, Yahoo! charges US $299 per web address for Express Inclusion
Service. For profit-making organisations, it is more or less the only way to
get included in the Yahoo! directory index. Apart from surfers looking for
information through the search engine directory browsing, the added benefit
of inclusion in the directory service is realized through improved page-
ranking in search engines. The search engines while ranking the pages
consider the quality of page based on the number and quality of linkages it
has from other web site. In that case a link to the web site from directory like
Yahoo! is treated as a Good authoritative link.
Pay For Inclusion
Pay For Inclusion revenue model applies mainly to crawler-based search
engines. Although, almost all crawler-based search engines offer free
inclusion in the search engine database once you submit the URL of your
website, there is no guarantee how quickly your website will be included in
the database. The Pay For Inclusion model guarantees that a crawler will visit
the website within a certain limited time frame and include it in the search
engines, database. The model guarantees inclusion in the database. However,
it does nothing about the ranking of your pages. In order to improve the
ranking of the website the website manager has to indulge in search engine
optimisation techniques. In that sense, this model differs from the Pay Per
Click model, where the website can attain a top position instantly by bidding
for the top position. The Pay For Inclusion model offers following benefits to
the websites:
It ensures a quick inclusion into the database usually within seven days
or so. For example, Yahoo! ensures that any web address paying the
inclusion fee is indexed within 72 hours. This is especially important for
start-up and new Internet businesses to get some visibility on the
Internet.
The fee-based inclusion also guarantees that the site remains included in
the database as long as you pay or your subscription lasts, irrespective of
algorithmic or other changes happening at the search engine database
site.
The paid for sites receive faster and repeated spidering. In the case of
Yahoo!, the periodicity of spidering is at an interval of 48 hours. The
faster periodicity of spidering offers ability to see the impact of search
engine optimisations quickly.
Although, Inkotomi pioneered the model, Yahoo!, Altavista, Askjeeves and
many more search engines have utilised the model for revenue generation.
Search Engine Optimisation
Search Engine Optimisation refers to the process of designing and tuning the
content, look and feel of the website in such a fashion that it will achieve
higher ranking for the right set of keywords and phrase. A search engine
optimised website by appearing at the top 20 listing of the web search result
page has the potential for attracting a larger base of interested customers. The
volume and profitability of the web business depends on the number of
customers that you are able to attract in visiting the website, and then convert
them into buyers. The search engine optimisation ensures that your web
business will be found by a web surfer and then a high percentage of those
visiting the website will be the customers, for whom the content of the
website is relevant. The first step in the process of search engine optimisation
involves the enhancing of the visibility of the website. This can be achieved
only if your website is listed in various search engines.
Adding Web Pages to Search Engines
As discussed earlier, search engines are an important business service
infrastructure for locating web-based businesses/service providers that meet
specific requirements. Consumers surfing a vast cyberspace rely on search
engines for identifying sites of their interest. The scope of the search is
limited only to those sites that have been indexed by search engines. Thus, it
is important for service providers/internet-based businesses that
websites/pages describing business activity should be part of search engines.
Most of the search engines perform crawler-based information collection and,
thus, eventually reach and index the website, if linked to other pages that
have already been indexed. However, indexing through this process may take
time. Thus, e-commerce businesses have to make special efforts to get
themselves indexed fast on as many search engines as possible. For this
purpose most search engines provide a mechanism for submitting website
URLs.
The submission is a simple process. The search pages of search engines
such as Google (http://www.google.com), and AltaVista
(http://www.altavista.com) have a clickable link to “Add/Submit a URL”.
Following through the process on the web page will allow submitting a URL
for indexing. The process may typically require information, such as
keywords, short description, author/owner of a page. This process can only
ensure a search engine listing, but the idea behind getting a listing in a search
engine is to appear in the result page in a relatively high position for
keywords phrase searches that are relevant to the website, and are likely to
result in a business opportunity/deal for the site. Thus, there are two concepts
that need to be clearly understood and distinguished. The first concept is
about getting listed in a search engine, and the second concept is about search
engine optimisation.
Listing in a search engine, as stated earlier, refers to the act of getting a
website listed with search engines. The term search engine registration is also
often used in this context. Getting a page listed in a search engine does not
mean that the page will necessarily rank among the top few for particular
keywords/phrases. It simply means that a search engine knows that a page
exists, and will form a portion of the cyberspace that is searched by the
engine.
The term “search engine optimisation” refers to the act of tailoring the
contents of a website, with the possible use of the meta tags described earlier,
so that it may rank reasonably high, to get a chance visit by the searcher of
particular terms.
The process of getting listed in crawler-based search engines, as described
above, is fairly simple. If a page has been linked by pages that are already
part of the crawler’s visited space, it will also get listed in due time. In
addition, web page writers may also utilise the “Add URL” facility offered by
most crawler-based search engines. The direct address of the “Add URL”
feature of some of these engines are as follows:
Google (Add Your URL Page) – http://www.google.com/addurl.html
HotBot (Add URL Page) – http://hotbot.lycos.com/addurl.asp
Fast/AllTheWeb (Add URL Page) –
http://www.alltheweb.com/add_url.php
AltaVista (Add URL Page) –
http://addurl.altavista.com/sites/addurl/newurl
Although, the above mechanism will get the site included in the database
of search engines, there is neither a guarantee about the time it will take nor
the frequency at which the content will be refreshed. For a commercial
website, this does not provide an acceptable option. After all, once in a
business, the website manager would like to ensure that the address starts
appearing in the search result pages for the relevant queries. Also, initially the
results may not appear high enough in the position so the manager may like
to fine-tune the content with search engine optimisation techniques. With no
guarantee on the refresh frequency, again the site manager will not be able to
see the impact of the optimisation in a fixed window of time. Many search
engine and directory service providers offer a payment based option (Pay For
Inclusion) to address these issues. As stated earlier, roughly over 70% of the
search traffic is aggregated by the top two engines, namely, Google and
Yahoo!. Thus, the site managers should make sure that they are included at
least in these two engines.
Inclusion in Google
Inclusion in Google search engine is free, even if you do nothing; the crawler
service of the Google may end up visiting your site by following a link from
some other site and include it in the database. Rather than waiting for it to
happen inevitably, it is recommended that the site manager should directly
submit the URL to Google website for inclusion. The site manager can
submit the information about the website to be indexed by Google through
the following URL:
http://www.google.com/addurl/
The simple procedure requires the website address, comments and
verification to distinguish between a manual versus machine submission.
Although, Google does not guarantee inclusion of every site submitted for a
variety of reasons including the inappropriateness of the content, it is a good
idea to submit the URL of your website and some internal pages that you
may deem important. In addition to this approach, Google also offers another
option for submitting a list of URLs or Sitemap file to the verified site
owners. This service usually results in a faster turnaround time for indexing
and is available free of charge. In the case of Google, once, you have
exercised any of the above options for submission, the crawler from the
Google will do the rest and index the other linked pages. The above
procedure only ensures inclusion in database, but does not address the
position of the page in search results. The Google uses a patented pagerank
algorithm for determining the rank of a page for a given query, and it depends
on various factors, including the number and quality of links that your page
receives.
Inclusion in Yahoo
Yahoo manages the website information in following two databases:
1. It maintains a Yahoo! search index of several billion pages whose
content is overwhelmingly (99% ) populated through the crawling
process. This service is available free.
2. It maintains a human-edited own directory of websites. The Yahoo
directory submission comes at a cost for all the commercial websites.
Yahoo supports both the paid and unpaid submission models, the paid
inclusion offers improved turnaround and time-bound actions. A site manager
can submit a website address for inclusion in the Yahoo! search index by
submitting the web site address through the follo-wing URL:
http://siteexplorer.search.yahoo.com/submit
For any website address submitted through the free submission
mechanism, the Yahoo! crawler will visit the site, extract other links and
discover the pages so far not discovered by the crawler in the past and add
them to the search index. The mechanism does not provide any time limit for
crawlers’ visit and also frequency periodic visit for refreshing the index with
the content of website not specified.
As an alternative, Yahoo! provides a paid inclusion option as well. In the
paid subscription model, the website manager is guaranteed that the crawler
will review the content in four days and the index will be refreshed by a
periodic visit of the crawler every seven days. All this is available for a cost
of US $49 per year for a single address. In addition to inclusion, Yahoo! also
provides analysis reports like for what keywords your site received click
through, it’s ranks for various search keywords etc. This information is
extremely useful in carrying out optimisation and monitoring the impact of
such efforts.
Inclusion in Yahoo! Directory is done only after a human editor reviews
the content of your website. For all the commercial websites, inclusion in the
directory service comes at a non-refundable cost of US $299. This guarantees
that your submission and content of the website will be reviewed with in
seven days. The fee does not ensure that the web address will be included, it
only assures a decision regarding acceptance or rejection is made within
seven days. For the accepted sites, the fee provides for one year inclusion, at
the end of the year your web address will be reevaluated for fee. For new
commercial sites that are trying to improve their ranking, the Yahoo
Directory inclusion may be well worth the cost. As during the initial period,
the website address may have very few links and that, too, of moderate-to-
low quality. The inclusion in Yahoo! Directory provides an authoritative link
of high quality to the website address, boosting its ranking.
A search engine’s listing, as stated earlier, only ensures that the web pages
will be part of the cyberspace that is searched by the engine for various
queries. With an astronomical increase in the number of sites, it is not
uncommon to see thousands of results for a simple query. Usually, the result
set of thousands of URLs is presented to the browser in a ranked order. Every
e-commerce business/service provider would like its site to appear ranked
high for a suitable query. The ranking of a document depends on a variety of
factors considered by search engines. Some of the essential factors have been
described in later section. Thus, it is important to understand the factors and
tailor the contents, the description, and keyword meta tags appropriately.
Some of the strategies that are commonly used are as follows:
Choosing the Right Keywords
Keyboards are the words that describe a site the best. They are determined by
visualising a search which will throw up this site at the top of the search
results page. For example, if the site being submitted contains information
regarding web surveys, or internet usage surveys, a person is searching for
information on internet usage surveys should see the website pages at the top
of the results set. In that case the keywords should be Internet services. The
target audience, in addition to the content, has an important role in
determining the keywords. It is advantageous to use multiple words as a
keyword as single words tend to find a very large set of matches. In the
example, if “internet usage survey” is used as the keyword it increases the
odds for appearing at the top of the result set, rather than if all these three
words were used as separate keywords. The word “internet” alone will have
tens of thousand matches.
Position the Keywords
The location of the keyword in web pages is crucial. Many search engines
pay heed to the position where the keyword appears on a page, during the
ranking process. Important target keyword appearance in the page title is
important. Many search engines would poorly rank even perfectly relevant
web pages, due to their failure to put target keywords in the page title. The
use of important keywords in the page headline, and high up on the page, is
weighed favourably by engines while ranking them. It is important that the
target keywords should appear in the first paragraph of a web page. Tables
should be included with caution in web pages. As table contents are viewed
by search engines one column at a time, a keyword appearance in the fourth
column of the first row will appear to be quite far down. The Javascript and
VBscript code in the beginning of a page also has the same effect of making
the keyword appearance lower than it is to the search engine.
Relevant Content
Irrespective of how the key words are chosen and positioned, search engines
are not likely to rank them high, if the pages do not contain a content that is
relevant to the keyword. Keywords should be reflected in the contents of the
pages. Many graphic intensive web pages may not have target keywords
appearing explicitly on the web page’s HTML text. The search engine will
skip the graphics content and will miss out on the relevance of the page. To
be on the safer side, adding HTML text with keywords in the main body of a
page in all such situations, makes the relevance of the page content to the
target keywords obvious to search engines, as well as users.
Avoid Search Engine Stumbling Blocks
Crawlers tend to access web page content similar to a text-oriented browser
like Lynx. It is very likely that many a crawler will skip the images, image
maps, and even frames. Thus, the content in such pages may not get indexed
properly. In order to ensure that web pages with image maps get indexed
appropriately, page designers should include HTML hyperlinks in an explicit
form, in addition to the image maps, as much of the relevant content is likely
to be in the linked pages rather than the home page.
The dynamic content generated through the CGI may also cause problems
in getting pages indexed properly with search engines. Most crawlers used by
search engines do not follow the CGI-generated dynamic pages. It is difficult
for crawlers to locate the content for such pages, for indexing purposes. To
avoid the problem, designers may consider putting up the first page, with
contents in it, for indexing purposes and then generating additional pages
with the CGI.
Many content-related problems that arise due to use of tables, scripts, and
CGI can be addressed through the use of appropriate meta tags. Description
meta tags can be used for providing a brief description of the web site, a
majority of engines support the use of description meta tags and take the
summary of a web page from the description meta tag.
Get Linked by Relevant Sites
Many websites try manipulating the ranking by placing the keywords and
words in the web pages. To overcome this, major search engines use link
analysis as a factor in ranking algorithms. The chances of getting good sites
to link to a website are few. Hence, analysing the links provides search
engines with a useful mechanism for evaluating the relevancy of the pages
for the given keywords and topics. Link analysis is not based simply on how
many sites link to the pages, but also evaluates the validity and relevance of
the links. In order to improve ranking based on the link analysis, the target
keywords and top ranked pages for these topic/keywords should be
determined. The administrators of these pages can be requested to link the
said page. Competitive sites may not agree to it, but some others may.
Getting links from these pages is likely to raise the ranking of a web page.
Additionally, since the linked pages are ranked high for the target keywords,
more visitors to these sites may follow the links and end up visiting the said
page as well.
Submit Your Key Pages
Although, search engines index all the pages that are linked to a web page
submitted for indexing, as a part of a recursive process of following the links,
it is better to submit two or three top level pages that best summarise the
website, as an insurance against search engines missing out on following up
some links.
Improving Searchability—Meta Tags
The search for a keyword or a phrase in cyberspace results in hundreds and
thousands of document URLs being returned. No user is in position to browse
through all on them. In all likelihood the first few URLs may be examined, to
locate documents of interest. It is important that the first few results shown,
for keyword/phrase searches, should be highly relevant.
Each search engine uses its own method for computing the relevance score
that is used for ranking, they are closely guarded trade secrets. However,
some general principles, which are borrowed from the text retrieval literature,
are discussed in the following paragraphs.
Text retrieval systems use frequency of the term, positioning of term in the
document, weighting, and proximity, as ranking criteria. Frequency of a term
refers to the number of times a term appears in a document. Documents in
which the term appears several times are ranked higher. The approach has a
serious flaw as a longer document may have the term appearing more often
than a shorter, but more relevant, document. This issue is addressed by using
the frequency of term relative to the total number of words in a document.
The importance of term positioning can be expressed with the example of
a journal paper. Journal papers have a title, abstracts, keywords, and the main
body. Generally speaking, a term appearing in the keywords has more weight
than one appearing in the title or abstract portions. The term appearing in the
body alone has lower relevance, when compared with others. Web pages also
have meta tags for describing the document, keywords, and title. Search
engines use the position of a term’s appearance for granting it higher
relevance i.e., if it appears in the following areas: title, the meta keywords,
meta description, first header, or first paragraph.
The other technique used in text retrieval is of term weighting. This refers
to the practice of making infrequently used terms, that do occur on pages,
more important than those which are common. Infrequent terms are given
more weight compared to the more common terms on the same pages.
Similarly, words that are extremely common, such as “and”, “not” etc., are
given zero weight during searching and ranking of documents. These words
are also often called stop words. Finally, in search queries with more than one
word, the proximity of words in the document also affects relevance scores.
Basically, the closer the positions of search terms, in a web page, the more
relevant they are considered to be.
Basic knowledge of the above can be used for improving the searchability
of a document for relevant search terms. Because, in electronic commerce it
is important not only to be ranked high enough to be visible to likely
customers, but is also far more important that on visiting site they find the
relevant information, service, or merchandise they are looking for. In this
regard while preparing a web page the issues of proximity, term weight, and
frequency should be kept in mind. The positioning part in web pages can be
addressed by paying attention to meta tags and crafting their content
carefully. The important meta tags in a web page are “description” and
“keywords”.
Meta tags are designed to be a useful mechanism for summarizing the web
page. Many search engines use this web page author defined summary for
indexing purposes, and place additional weight on the terms that appear in
these tags. In many cases the page designers may have the starting page
loaded with graphics and image-maps. Such pages have little textual
information that can be used by search engines. In these cases, meta tags such
as description and keywords can be used for describing the page content.
These tags appear in the <head> section of a HTML document. The
following example illustrates the syntax and use of meta tags in a HTML
document.
<HEAD>
<TITLE>My Personal Page</TITLE>
<meta name=“description” content= “Internet Commerce Research
Center,E-commerce Resources, Research on models of electronic Commerce,
Network Infrconsidered to be a structure,EDI,Web,E-Commerce in India”>
<meta name=“Keywords” content=“ Internet Commerce, E-Commerce,
India, IIML Web Usage Survey,Web Databases,EDI,Agent-based
Ecommerce,E-Commerce in India”>
</HEAD>
The meta tag description is used by search engines for indexing purposes,
in addition, the search engine uses the description for summarizing the
content of the page. If a web site appears in the result set of some search, the
search engine will describe the summary of the page using the content of the
description tag. In absence of this tag the search engine may include the first
few words from the contents of the web site, for the summary, which may not
appropriately describe the content and intent of the web site. The meta
keywords tag provides the page writer a chance to categorize web pages,
using the keywords. In case of a keywords search, the web page is likely to
come up with some links, in the result set if then contain any of the words
listed in the keywords tag. For example, someone might enter “Web Usage
Survey” which matches with one of the keywords in the tag described above.
If the phrase “Web Usage Survey” does not appear in the contents of the page
as it is, without that tag, there would be no chance at all for it to come up.
It is important to remember, that these tags help in compensating for the
lack of text on the pages, and classifying the page contents by keywords.
There is no way to anticipate every keyword variation a person might enter
into a search engine. Thus, it is good idea to include as many variations of it
as possible, but it helps only to a limited extent. Meta tags are a tool that help
in getting around the aforementioned problems.
SUMMARY
With a vast amount of information resources available on the internet,
traditional surfing, or the word of mouth model, for locating any information,
does not scale up. For the continued growth of the electronic commerce, a
business service infrastructure that assists in the search and location of the
right kind of information and scales up well, is required. Directory services
and search engines have been providing this service for a period of time, and
have scaled up well. This chapter describes these two models of information
organization, their salient features, and applicability. The search engines have
evolved beyond the banner advertisement model for revenue generation and
have been willing participants in the Search Engine Marketing programmes
to enhance the revenue streams. The Pay For Consideration, Pay For
Inclusion and Pay Per Click are the current means of revenue generation.
These models offer an equitable and measurable mechanism to both the
website managers and the search engines for generation of revenue that is
based on performance. The information stored in search engines can be made
more meaningful in order to get the right kind of matches or those ranked
higher, for target keywords/phrases. Various tools and techniques that can be
used for this purpose have been discussed in this chapter. In order to search
for information, it is important for the user to formulate the right search
strategy. The issues involved in formulating good search strategies are also
discussed in this chapter.
REVIEW QUESTIONS
1. Describe the organization of a subject directory in a search engine like
Yahoo!
2. What is full text search engine?
3. Describe the use of meta tag keywords.
4. What factors are generally considered by search engines in trying to
determine the ranking of a page?
5. Describe the mechanism used by search engines for automated
collection of information.
6. Describe the strategy used by designers of web sites for getting a page
added in search engines, and getting it ranked high for target keywords.
7. Why it is important to formulate a good search strategy for locating
information of interest? Describe some of the factors that play a role in
search strategy formulation.
8. What is the difference between paid and pure search result pages?
9. Describe the role of search engine optisation in generating traffic to a
website?
10. Distinguish and contrast between the Pay Per Click and Pay For
Inclusion models.
11. How does the Pay For Consideration in directory impacts the ranking
performance of web pages on crawler-based search engines?
With the growth in the number of users, the internet is increasingly seen as a
commercial medium with immense potential for information sharing, market
transactions, advertising, and promotions. Many internet service providers
(ISPs) now offer internet connectivity to the masses, and this is changing the
profile of the users on the internet. The growth of information content
providers, such as Newspapers, Magazines, and Electronic Newsletters has
mirrored the growth of internet users. The entry point and the cost of
publishing being minimal, many new publications, with a wide audience
reach, have only accelerated the process.
With the growth of advertising on the internet, revenues crossing the US
$21 billion per year mark in 2007, new publication models have begun to
find a commercial footing. At present, however, the major source of direct
income is from advertising. As the shift to digital economy continues, and
access to content and material on a chargeable basis becomes viable,
advertising on the internet will change and mature.
The interactive nature of advertising on the internet increases the control
of the information receiver over the information they are exposed to. The
consumer, given the option, is likely to be more selective in defining the kind
of information she is willing to receive, interactivity gives them that option.
Unlike mass media such as newspapers, radio, and television, the internet
audience is not a captive one and thus advertisers have to work harder to
entice them. For the advertising agencies, advertisers, and developers of the
new media, it poses newer challenges in the form of a non captive audience,
and an information rich and savvy means to guide consumers towards the
information source, in a user friendly manner.
With the growth of traffic in the electronic marketplace, for some
advertisers, the internet will prove valuable, but for others it will be an
expensive failure. The reasons for their failure or success are discussed in the
subsequent paragraph of this chapter.
INTERNET ADVERTISING
Select newsgroups, email messaging, and some list servers have been utilized
for commercial messages and advertising purposes, in a protracted form for
quite some time in. The emergence of the World Wide Web architecture, and
its ability to deliver animated multimedia content online, is by far the most
appealing part of the internet structure, for advertisers. Advertisers can
directly relate to the multimedia aspect of the web content as it is more
closely related to the types of advertisements with which they are familiar—
color spreads in magazines, boxed advertisements in newspapers, and
commercial spots on television and radio.
The objective of advertising is to increase the awareness about the
advertised product, program or service, ultimately translating into an increase
in the sales volume or activity, by supporting a thought out and articulated
marketing program. Thus, each effective marketing program requires an
appropriate advertisement suited for the mission, the message, and the target
audience. The success of a good market campaign lies in the ability to discern
two essential elements of this target market: (1) who the audience and (2)
what is their buying pattern.
Advertising supports the marketing program by influencing, through
impressions, the “audience”. An audience is that part of the target market that
can be expected to experience the advertisement or series of advertisements
(a ‘schedule’). It is a well-defined and measurably quantifiable subset of the
target market. The nature of the audience mix in terms of demographics,
psychographics and other factors determine the ‘composition’ of the
“audience”. The size of the “audience”, relative to the target market, is
referred to as the advertisement’s ‘reach’. Every time, a member of its
audience experiences an advertisement, it is said to have made an
‘impression’. These impressions must be effective. The effectiveness of an
advertisement is usually measured by recall, i.e., can a member of the
advertisement’s audience remember the advertisement at a later date?
Many an advertisements needs to be seen several times before it can be
recalled. The number of times a member of the audience must be exposed to
an advertisement before it can be recalled is referred to as the “effective
frequency” of the advertisement. The effective recall frequency falls
somewhere between four and seven, for most traditional media advertising.
Exceeding the effective frequency is unnecessary. It might even be
harmful in some cases, because the advertisement may then simply become
part of the general background or scenery and subsequently get ignored. An
advertisement’s effectiveness tends to follow a normal distribution or a bell-
shaped curve, with one to three low exposures, four to seven as high, and
thereafter the curve tails off again.
In traditional media, advertising costs are not linked directly to
effectiveness, but rather are most often determined solely by the number of
impressions that a given publication or site can deliver to its target audience.
These rates are usually quoted as cost per thousand impressions a (CPM).
The emergence of the internet as an information exchange and
communication medium, through FTP applications such as Archie, Gopher
and Veronica, Bulletin Board Services, electronic mail facility, has opened up
new avenues for advertising. Advertising, through these media, has been in
existence for a decade. It is the emergence of the world wide web, powered
by HTTP and HTML, with multimedia publishing capabilities, that has made
it a means for mass communication.
The fundamental building block of web advertising is the sponsored page
itself. Web users downloading a popular page would be presented with the
sponsor’s advertisement, in a passive manner—the advertisement requires no
interaction or activity from the user. Early advertising models simply
involved the advertiser paying the web page owner/publisher, on the basis of
the page’s popularity; either a monthly fixed fee, or more often an
impressions fee based, similar to the CPM in traditional media advertising.
A particularly useful aspect of the web is its ability to engage the user in a
more active marketing message, than can be achieved through the passive
‘witnessing’ of a magazine spread. The advertiser’s objectives, therefore,
gradually shifted from exposing the user to simple impressions, to enticing
the user into visiting the corporate web site of the advertiser, for a more
complete marketing dialogue.
The web operates on a referral mechanism, as people have to know the
address of a page to visit it. This can be accomplished using traditional media
mechanism for building traffic to the page, offering dynamic contents of
value, and getting it indexed in various search engine databases under
appropriate categories. The strategy of just build it well and they’ll find you
is not appropriate to the information rich internet environment. Web
advertisers are crucially dependent on links to their sites, which have evolved
from the simple logos of early sponsoring into what are now called ‘banners’.
The most obvious application of advertising skill in the web is now in the
creation, placement, and operation of these ‘active advertisements’.
Although the internet offers a huge, unlimited global advertising
opportunity, there is some need for caution. The common myth about the
internet offering global coverage, is certainly true, as the number of internet
users are several tens of millions, and the demographic mix is appropriate for
certain classes of products and services. However, it would be unwise to
assume that this counts as a ‘global audience’ for any advertisement on the
internet. An internet advertising campaign for certain goods will attract
attention only from a subset of the global audience, and is limited to those
who:
Know of its existence,
Are interested in the products and services, and
Intend to receive the commercial message itself.
Some products and services are definitely of interest to almost all internet
and web users. For example, by definition, internet users are potential
customers for computer hardware and software products. The internet
population, therefore, is an almost ideal target market for companies such as
Microsoft, Adobe, Dell Computers, and Netscape. The relationship between
the target market and the internet user community may not be so well defined
in some other products and services. For example, the relationship between
Ford’s target market and internet users is not very obvious. For others, it may
just be a palpable belief that their products can receive global exposure
through a corporate web site. Had they spent the money on more traditional
media, it would have cost much more then the expense of establishing a few
web pages. However, low costs alone are not enough to make web
advertising viable. A strong presence, which translates into considerable
audience reach, is required for successful internet advertising. Many
companies, from specialty manufacturers to global trading concerns, have
found that the internet is definitely not the gold mine that it was hoped to be.
In fact, about 40% of corporate web sites, built with the objective of global
marketing presence, were abandoned as a result of disappointing internet
visibility and overall returns.
The other aspect of the global media myth is that like the audience of a
television program, the internet audience is a passive one. Early internet
advertisements mimicked this, assuming that a sufficiently captivating image
would hold the viewers’ attention long enough for the product message to be
transmitted. On the contrary, the internet user is not recumbent. Internet
surfing is an active phenomenon, with users rapidly hopping from one
internet site to the other. The challenge for the advertisers therefore is not
simply attracting the viewer, but also captivating and retaining him.
Advertisers need to entice the internet users to forums, or identify
appropriate forums, such as chat rooms, mailing lists, bulletin board services,
newsgroups, FTP archives and web sites. Once attractive forums have been
identified, the advertiser using internet media has several advantages over the
regular media advertiser. This is primarily due to the fact that users have a
choice in deciding to visit and spend time over the advertisement. Thus, any
viewer who spends time is likely to be an interested one. Internet users
themselves evolve, from being novices to becoming more sophisticated, in
using and trusting electronic gadgets and the internet. Three new market
segments seem to have clearly emerged on the horizon. We classify them as:
1. Net-Surfers: This segment consists of new internet users, usually young
persons with short attention spans. These people tend to hop from site to
site, usually trying to discover more and more; if something looks
interesting at a site, they may scan it, or download it otherwise move on
to next site. People in this segment may be browsing several
documents/sites simultaneously. It is the segment that may be very hard
to appeal to, but is attractive to marketers and advertisers. The decisions
made by this segment tend to be impulsive, and buying right off the net
is quite common.
2. Net-Buyers: This segment of users spend a lot of time online as a part of
their business activity, usually at their workplace. It tends to be
dominated by software professionals, academicians, researchers,
engineers, and others employed in the online service provider industry.
According to first IIML Web usage survey, 1999, this group consists of
nearly 60% of the internet user population in India.
3. Net-Consumers: This segment consists of users who access the network
from their homes. It represents families, and offers the opportunities for
the retail industry, entertainment industry, and convenience stores. It is
the segment that holds the biggest promise, as we move towards a digital
economy. Advertisers and marketers can influence this segment by
making it more convenient to shop online rather than visiting local
stores. In India, this segment already constitutes 30% of the internet
users.
There are two ways by which the internet user can frequent a forum: by
typing the forum address explicitly into the client-program, such as a web-
browser, or by using a referenced link from the current forum.
There are several ways to tell a user about a link to a web page. The most
obvious way is to have the site can be listed by one, or all, of the web search
engines. In addition, external marketing and advertising, that forms a part of
the broader program or campaign, can also reference the web site. Many
newspapers, television or magazine advertisements now include URL’s for
advertiser’s home pages. Finally, business cards, letter heads, exhibition
boards, product wrappers, etc., can all carry the URL alongside the logo.
SPONSORING CONTENT
The banner, not being part of the web surfers search pattern, tends to get
ignored, unless the message is directly related to the surfers, intended
content. Therefore, successful web advertisers must ensure that their content
—commercial messages and enticements-is included as part of the user’s
search and surf patterns, rather than as a separate, free-standing and easily
ignored part.
Perhaps, the simplest and most obvious model for this is ‘product
placement’. A sponsor’s product—soft drink, a motor vehicle, clothes etc.—
is used and presented in a blatant and explicit manner within the film,
television show or novel. The application of this approach to the Web Pages
is easy to see. At the simplest level it is called “content co-branding”, the
sponsor’s messages can be woven in throughout the content of a sponsored
web page, e.g., a sponsorship deal between a golf equipment manufacturer
and the web site covering sports information. Content weave sometimes leads
to a compromise in site quality, through over promotion of the sponsor’s
interest. A better alternative to content weaving is the use of ‘microsites’. The
idea behind a microsite is that the sponsor funds or provides a smaller set of
pages—much smaller than the primary corporate pages—that are of
immediate and specific interest to the sponsored site’s visitor. Usually these
are associated with ‘infotainment’ sites such as online web magazines, where
the microsite acts almost like a newspaper insert. These microsites have
sometimes been called ‘brand modules’ or even ‘cuckoos’, since they are like
eggs placed in another bird’s nest. The important point is that these
microsites are developed specifically to follow their basic structure,
presentational feel, and to be intimately embedded with the core content,
without compromising it.
The microsite can make it clear that a set of pages is sponsored, or can
choose to disguise the fact; it can even include an explicit link to the
sponsor’s site, for those interested in more information about the particular
brand. These microsites have been used by a variety of successful advertisers;
VISA, for example, sponsored a collection of such pages within Yahoo!.
The Sponsorship Process
While the microsite, or even more intimate sponsorship, is more likely to
succeed than simple banner links, there is still the question of ensuring that
the sponsorship deal itself is successful. In making a formal sponsorship
arrangement, the sponsor must ensure that a wide variety of contractual
conditions are put in place. These include, where the links to the sponsored
content will appear, the guarantees a web site owner provides against system
failures, an agreement not to carry rival products, and establishing the
responsibility for maintaining and updating the microsite pages. It is also
necessary to establish the conditions under which the sponsored site will gain
additional revenue. In electronic commerce, there may be a case where as a
result of a link from a sponsored site, products are sold. In this situation, it
would be entirely appropriate to reward the sponsored site for its
effectiveness. For example, Amazon offers a commission sites providing
links to its online bookstore, when links result in a sale.
Because of practical considerations, advertisers are looking for
alternatives to the interactive medium of banners and web sites. This has
resulted in a return to the traditional ‘push’ form of broadcast advertising,
within the web.
INTERSTITIALS
In 1997, Berkeley Systems introduced a new model of serving online
advertisements. These advertisements, referred to as interstitials, appear in
between on screen activities, such as pushing a button, transition of the
screen, in game shows, or in interactive session situations when you reach or
cross certain thresholds. One of the early uses of the interstitial was in the
“You Don’t Know Jack-NetShow”, where after every five or six questions in
an interactive game, a mini-commercial, with rich multi-media, capability
popped up. Due to rich media content ( audio, video and images etc.), and
being integrated as a part of the game, the model has been more effective
compared to traditional models. In the online advertising scenario,
interestitials offer more creative advertisements compared to banners. A
judicious mix of audio, video, and images can render a television-like
advertisement over the internet. This television-like advertisement quality of
interstitials captures the users, attention actively, unlike the banner that may
simply be ignored by them. Given the present bandwidth bottlenecks of the
internet, at times it may not be possible to deliver these rich media clips
online. Intermittent and jerky delivery over the network may actually provide
an experience worse than a banner.
SUPERSTITIALS
Another alternative model, introduced in May 2000, addresses many of the
problems faced by the interstitials. Superstitials provide the opportunity to
create larger and more creative online advertisements, using a slightly
different delivery mechanism, that addresses the problem of degradation in
user experience, at the time of rendering. This model, like interstitials,
overcomes the creative limits imposed by the banner’s position and size.
Unlike interstitials, that suffer from degraded user experience problem due to
online delivery limitations, superstitials use cache-and-play paradigm for the
delivery of advertisements. The superstitial model does not interfere with
web site content loading. Instead, once all the content has been loaded and
the user is browsing the information, the superstitials are cached into the
browser’s cache, in the background. These advertisements are played once
the content has been fully loaded the and the user decides to move to another
page. At the time of transition, the advertisement appears in another window
and starts playing from the browser’s cache. The rendering of the content is
not effected or slowed down as the whole multimedia content has been
downloaded in the cache. During the rendering of the superstitial, the
transition page gets downloaded from the network. Thus, in this model, the
advertisement never competes for the bandwidth with the web content.
Advertisements in this model play during the transition, triggered by a
mouse click, and capture user attention for a period. Unlike banner
advertisements that can be completely ignored by the user and yet get
counted for payment purposes, here advertisements that have been fully
downloaded are the only one that get counted. In the model, the user either
get to see the fully downloaded advertisement or nothing at all, in fact the
user is not even aware that an attempt was made to download an
advertisement. Thus, advertisers have complete control over the count of
impressions delivered to users; this gives them freedom to create compelling
advertisements that motivate the user.
OPT-INs
This is an e-mail based advertising technique where users explicitly opt to
receive advertisements. The opt-in e-mail contains information or advertising
regarding products or services that users have requested to receive, during
some form fill out process. In this advertisement model, a web site attracts its
visitors to register for some services, such as a free web mail services,
competitions etc., and requests them fill out registration forms for the same
purpose. The forms also contain information/options, identifying many
subject or product categories that may be of interest to users. At the time of
filling out the forms users may tick/opt to receive information regarding some
or many of the categories in which they are interested. With the emergence of
newer technologies in receiving e-mail, opt-ins can be received through e-
mail, through PDAs, mobile phones, and pagers.
Pop-Up and Pop-Under
The pop-up advertisements utilised by many a website is experienced quite
frequently. In this advertisement model, when you are visiting a page of the
website containing a pop-up advertisement, a separate window “pops up” and
the advertisement is displayed in this window. A user in most of the
situations has to move the pop-up window either out of the way or has to
close it in order to focus on the content of the website. Thus, many people
feel highly annoyed by the pop-up advertisement. Most of the current
browsers support the pop-up blocking, and hence many users enable the pop-
up blockers in their browsers. The pop under advertisements are slighlt less
intrusive as these advertisements hide themselves under the content of the
web page. They appear only when a user is trying to browse through the
specific content and are therefore are less intrusive. Despite the annoyance
factor experienced by many users, studies have shown that these
advertisements are far more effective than banner advertisements. A typical
banner advertisement may be able to get a 0.2 to 0.5% click-though rate or, in
other words, a 2 to 5 click-throughs for every 1000 impressions, The pop-up
advertisements, during the first few exposures, have been able to achieve
around 3% click-through rates i.e., 30 click-throughs for every 1000
impressions. Consequently, despite the perceived annoyance factors websites
use them quite often and advertisers also pay more for pop-up and pop-under
advertisements. The common going rates of pop-ups and pop-under
advertisements on a website is 4 to 10 times more than that of a banner
advertisement.
Floating Advertisement
The floating advertisements are created, as the name suggests, for remaining
visible in the viewing area of the browser window for a specified time. The
time typically varies from 10–30 seconds. Most of the time, these
advertisements place themselves on top of the content of the page that you
are trying to view and thus, grab your exclusive attention. Some of these
advertisements may have an escape, such as “close” button for the users,
while the some may even follow your mouse movement. As these
advertisements, like a television advertisement grab the screen by appearing
on top of the content, thus interrupting the activity that a user was engaged in.
These advertisements consist of informational text and pictures, interactive
content or flash content that may capture the entire screen for a few seconds.
As a result, users cannot ignore them and hence from the branding point of
view, they are far more effective than simple banners, customized banners
and sidebar advertisements. Also, a well-designed campaign utilising the
floating advertisements can be highly effective, and, as per various Internet
advertising associations data, can attain a click-through rate as high as 3%,
i.e., 30 click-throughs per 1000 impressions. The enhanced branding ability,
coupled with higher click through rates, have made them a popular medium
for advertisements. Since, these advertisements fetch more revenue,
anywhere from US $ 3 to US $30 per 1000 impressions, the various websites
and portals are willing to run them at the cost of annoyance caused to the
visitors of the websites.
Unicast Advertisements
The unicast advertisements are basically the reincarnation of television
advertisements in the Internet environment. These advertisements are
animated and have sound and run like a television commercial in a separate
window. The typical advertisement has a run length of anywhere between 10
and 30 seconds. The unicast advertisements have an additional advantage
over the television commercials, the user can click anytime on the
advertisement and access additional information. According to the Internet
Advertising Bureau, the unicast advertisements have been able to achieve as
high as 5% click through rates, i.e., 50 click-throughs for every 1000
impressions. Due to higher click-through rates, these advertisements are able
to fetch a lot more revenue to the website. The typical rates for running 1000
impressions of these advertisements are in the range of US $30.
SUMMARY
In conclusion, advertising on the internet is relatively cheap, covers a
widespread audience, and provides exciting opportunities of exploring a new
and interactive medium. The emergence of internet advertising is likely to
have wider implications for businesses, than many imagine. Its effects will
not be confined to the online world, but will extend to traditional marketing
activities and processes too. Internet advertising holds many opportunities
and risks, but for those who rise to the challenge, it will more than justify the
efforts required. Several advertising models that have been effectively
utilized on the internet are discussed in this chapter. Banner advertising has
been the most widely deployed model. This chapter deals with the basic
approach of the model, payment and effectiveness, and customized banner
delivery. The chapter also describes other important internet advertising
models such as sponsored content, screensaver, push broadcasting, corporate
web sites, interstitials and superstitials. The chapter also discusses the
weaknesses related to measurement discrepancies and metrics. Finally, the
chapter discusses the various pricing standards that are prevalent in
advertising on the Internet and some important online placement forums that
are available to the advertisers.
REVIEW QUESTIONS
1. What is a one way advertising channel?
2. How does internet advertising offer a two way channel?
3. Describe banner advertising and related pricing models.
4. What is the push broadcast advertising model?
5. What is click-through rate? Discuss its importance in internet
advertising.
6. Describe the measurement problem in internet advertising.
7. What is the placement network and how does it operate?
8. Discuss the pros and cons of Cost Per Acquisition model for revenue
generation in advertising?
9. Why are search engines an effective platform for advertisement
placement?
10. What kind of advertisements are better suited for vertical portals?
The growth in the number of mobile telephone service users in the past ten
years has well surpassed what took plain old telephone service 50 years to
achieve. At the dawn of 2005 India alone had 40 million mobile telephone
subscribers. By 2006, the numbers of mobile devices are expected to cross
the billion mark. The digital revolution sweeping the world today is being
further fueled by fast paced innovations in electronic and wireless
technologies. The wireless electronic devices originally used for voice
communication were limited by the available bandwidth and could barely
achieve the data rates required for any reasonable application. But, with the
innovations in the past five years in terms of transmission mechanisms and
data rate, multi-media messaging (MMS), face-to-face communication, video
transmission through mobile communication devices have become available
to users of 3G. With the through emergence of 3G, mobile devices are
becoming a central part of people’s life today. Before we proceed further let
us understand 3G that seems to be making it all possible. 3G refers to the
third generation of wireless communication technologies that enable high
speed data access (commonly up to 2Mbps) over wireless networks. It is
important to note that 3G refers more to a range of data access speed rather
than any particular technology.
Mobile electronic devices operating over wireless networks with data rates
of 2 Mbps offer altogether new ways of conducting business. Over the past
decade, advances in information technology have been leading innovations in
business model design and strategic direction. Technology has come to
occupy a central spot, not only in operations but in the strategy as well.
Companies, such as General Electric, who recognized the potential of
electronic technologies and the Internet are able to lead the innovation in
business processes and models, in addition to operation efficiency. Wireless
communication technology with fast growing achievable data rates are a new
and important frontier to watch out for. As these technologies hold potential
for increasing the reach and scope of existing business applications and
processess, they often offer alternate innovative business processe leading to
means for cost cutting, enhanced productivity and improved, efficiency.
In the global economy, keeping track of technological advancement has
become an arduous task as competing innovations continue to happen,
leading to a plethora of technology directions. The lack of any standards
although, desirable at the early phase of emerging technology, makes the task
all the more difficult. For existing companies with traditional strategies,
competitive forces surface from non-traditional sources. With Internet
penetration, seamless access through the world wide web gave rise to new
marketplaces and forced traditional companies to adopt electronic commerce
for their survival and growth. Today, the personal computer revolution of the
80’s has acquired the proportions of point of no return, and Internet
connectivity and digital transformation has advanced economy to the era of
electronic commerce and business. As the current advances in digital
transformations the lead the electronic commerce era are being cemented, a
new force due to innvoation taking place in the wireless technologies has
already begun to push organizations further up the ladder of evolution.
Mobility is the new buzzword; and innovations in wireless technologies are
the key drivers. As in electronic commerce, customers are no longer required
to reach out to their computers. Mobile devices are enabling them to access
the information, make bill payments, make reservations; play games,
download music, and videos, interact with friends, family or relatives; and
carry out a transaction on a small screen of mobile devices. Access to the
electronic marketplace has been freed of all strings. The existing electronic
commerce strategy and processes may not suit the smaller screen space, the
mobile connectivity with still slower bandwidth, location specificity of the
requirements.
Advances in transmission technology and standards in mobile
communication systems have made it possible to achieve transfer rates of 2
Mbps over wireless networks. A single channel GSM/TDMA system can
ensure a 14.4 Kbps transfer rate. The convergence of mobile communication
devices with Internet content is an inevitable area of growth. The inherent
advantage of lower entry and deployment costs, ease, speed of deployment,
and demand-based expansion will continue to fuel the growth of the number
of users. Compare this with the wired world saddled with the right of way,
laying of physical cable issues, and competition from entrenched local
exchange carriers. In countries like India, the on demand availability of
service with all added benefits the mobile access has already seen a strong
growth. As the sophistication of mobile communication services continues to
grow with Short Messaging Service (SMS), CPDS, and other related message
and data packet services, the expand devices with embedded processors in
them can be used for more than conversations. On the other hand, as the shift
to digital economy continues at an unprecedented pace, the demand for
content, be it stock market updates, personal banking information, digital
diaries, and other information, so readily accessible and available from the
Internet, is likely to grow exceedingly. According to Metcalfe’s law, the
value of any network is said to be proportional to the square of the number of
its connected users. The growth of Internet led digital economy will be
directly beneficial, if the users of mobile devices are able to access and
transact on the Internet through mobile devices or personal digital assistants
(PDA) such as palmtops, mobile phones and pagers. The convergence of the
wireless world and the Internet is the next frontier that will complement and
fuel mutually beneficial growth in both sectors.
SUMMARY
The past decade was marked with an unprecedented growth in the number of
mobile phone subscribers. The voice only mobile phone network acquired
data transmission capability and devices are able to handle multimedia
content with ease today. Improvements in the available bandwidth for
accessing the digital content has provided and added a new dimension to the
online digital marketplace, and has given rise to the phenomenon commonly
referred to as mobile commerce. In this chapter, we discussed the reasons for
the growth of mobile commerce and the issues that still need to be scaled in
order to realize the full potential of electronic commerce. As mobile
commerce builds further and leverages on electronic commerce in addition to
creating newer opportunities, we described a mobile commerce framework
consisting of the following architectural elements.
Wireless network infrastructure
Information distribution over wireless networks
Multimedia content publishing technology for mobile devices
Mobile security and encryption
Payment services in the mobile environment
Business services infrastructure public policy, and legal infrastructure
Mobile commerce applications
For continued growth and success of mobile commerce in the
marketplace, it is important that all these elements are in place and the basic
issues related to these elements are addressed as well.
REVIEW QUESTIONS
1. What is mobile commerce?
2. Describe any four major advantages offered by the mobile commerce
environment as compared to commerce over a wired network.
3. List and explain the major impediments faced by the mobile commerce
environment.
4. Define the architectural framework of mobile commerce.
5. Describe and distinguish between 1G, 2G, 2.5G, and 3G mobile
networks.
6. What is the minimum functionality required of a base station in a
AMPS?
7. What do you understand by “Global System for Mobile
Communication”?
8. List and describe the functions of major subsystems of a GSM network.
9. List the databases maintained by the MSC in a GSM network and
describe the purpose of each of these databases.
10. What is GPRS? How does it achieve the higher data rates?
11. Compare and contrast EDGE with GPRS.
12. What is 3G network? Describe the five ITU approved 3G standards.
13. What do you understand by Wireless Access Protocol. Describe the
purpose of the WAP Gateway?
14. What are important layers in WAP? Describe the function of each of
these layers.
15. What is i-Mode service?
16. Why do we need mobile device compatible languages for publishing
content for handheld devices? Give two examples of such languages.
17. What are the four fundamental security issues in commerce and what is
the impact of the mobile commerce environment on these issues?
18. What are the online payment issues in mobile commerce?
19. What type of payment models have been used in mobile commerce?
20. Define issuer-centric and mobile network operator-centric payment
models.
21. What are the important revenue streams on which sustainable businesses
can be built in the mobile commerce environment?
22. Describe the additional benefits offered by location and search services
in the mobile commerce environment as compared to the wired network
environment.
TYPES OF AGENTS
Software agents can be classified on the basis of three orthogonal
dimensions; these dimensions are mobility, intelligence and autonomy.
Mobility
Mobility refers to the degree to which an agent can move through networks.
In some cases, the agent program executes only on the host system and does
not move at all, it interacts with other systems only through the
communication mechanism. In other cases, the agent program can be moved
to another computer and executed there. In still others, agent programs are
capable of suspending their own execution on the current system, move their
programs and execution state to another computer, and resume execution
there on the new system.
Static agents: These agents execute on systems in which they begin
execution, and interact with other systems using communication
mechanisms. They use embedded knowledge to assist in filtering and
processing volumes of information.
Mobile agents: These agents are not bound to the system in which they
start execution, and can, therefore, travel among other hosts in a
network. They carry out transactions without continuous, instructions
from the user. Mobility is achieved by transportation of state and code to
the new execution environment at the destination. Mobile agents are
well suited for e-commerce, as commercial transactions may require real
time access to remote resources such as stock quotes and agent-agent
negotiations.
Intelligence
Intelligence refers to the degree to which an agent can identify and perform
tasks, in order to meet the objectives specified by users. Intelligence is a
unique human trait. Human intelligence has several levels—the highest
probably being creativity. But, on the other hand, the mind has learned to
process constant sensory input signals at a trivial level so that the signals do
not overwhelm us. These signals are classified into manageable sets of
information by identifying the patterns, similaries, and differences. The
sensory input is put in an appropriate compartment based on a set of learned
rules. Each compartment has rules, for example, the human mind knows that
if it has wings and feathers, it’s a bird, if car races towards you, move to the
side, if you want to get the past tense of a verb not ending with “e” add “ed”,
otherwise add “d”. Now, all one needs to note is similarity, differences, and
changes. In case of exceptions, add the exceptions as well to the set rules.
Computers are faithful in following rules, if we teach them a set of domain
rules, they can follow them easily. The problem of teaching them how to
identify patterns is more complex, but it is possible to implement it in a
limited way. Agents vary in the degree of intelligence embedded in them. The
simplest forms follow pre-defined scripted paths; while more advanced
agents are driven by a set of specified rules, for a given problem domain. The
most advanced agents, or truly intelligent agents, are capable of observing
and learning from the environment training data set. These agents apply the
learned behavior on new situations, to meet the objectives of the users.
Autonomy
This refers to the degree to the which an agent can exercise control over its
own actions and state. In other words, autonomy refers to the agents’ ability
to act without supervision. In a typical user and computer program interaction
paradigms, programs act only when users initiate them to do so. In an
alternative to this paradigm, both users and computers can initiate actions and
monitor events to meet a set objective. The software that provides this kind of
human-computer collaboration for meeting goals is called an autonomous
agent. Autonomous agents utilize the knowledge gathered about needs and
preferences through past repetitive tasks, to assists in similar tasks. The
concept of autonomy is highly related to the concept of proactive behavior. It
emphasizes that agents do not simply act in response to certain changes in
input or environment, but, display goal directed behavior by taking the
initiative. This proactive behavior is a key element of autonomy.
The simplest agents interact with databases, applications, and services, in
order to determine alternatives that may meet the user’s objectives. The more
sophisticated agents may even collaborate and negotiate with one another to
meet the goals set for them.
Characteristics of Agents
are be programmed to perform tasks that meet the user requirements, based
on all available information and learned behavior. In the process of meeting
the goals set by the user, the agent may use its mobility, autonomy, and
intelligence, to the available limits. In the case of mobile agents, the agent
may start execution on remote systems and use the resources of the remote
systems. Resource usage of agents must be monitored to assure that they do
not use disproportionate resources and also because the remote systems have
to be assured that the safety of the system will not be compromised as a result
of agent execution. Agents must have the capability to find the resources that
they need. Also, agents must not divulge more information that they should,
while interacting with other agents. The key characteristics of agents include:
Agent Independence: Agents must be capable of providing the required
services without the user’s guidance or presence, when the conditions
are all met.
Agent Cooperation: Agents must communicate with each other by
exchanging messages in standard communication languages, and must
also cooperate with each other. Often, to perform complex tasks, static
and mobile agents must work together.
Agent Learning: Agents must have the intelligence and ability to learn
from their experiences. The adaptive functionality of agents requires
them to possess the characteristics of noticing, interpreting, and
responding.
Agent Reasoning: Agents must have the capacity to operate with
decision-making capabilities in complex situations.
Agent Interface: Agents must be able to encapsulate the operations and
data, and decuple them from interfaces with other agents.
Anthropomorphic interfaces can also be used to build trust, and make
the user comfortable with agents.
Software agents can be categorized on a spectrum, one end of which is
dominated by agents that simply mimic user actions when invoked. On the
other end are agents that can learn adoptively and use historical information
to draw inferences on expected behavior.
End User Taxonomy
From the perspective of application to the end-users; agents can be classified
in the following categories.
Desktop Agents
Operating System Agents: Interface agents that provide user assistance
in the desktop operating system environment. The user working in a
operating system environment may be trying to achieve a task, may have
forgotten all the steps. These agents observe the user behavior and offer
assistance that may lead to users accomplishing the task with ease.
Application Agents: Interface agents that provide assistance to the user
in a particular application. These agents operate in much the same way
as described above, but within an specific application environment. The
application agent in its simplest form can be seen in the MS Word
application; where a user, trying to accomplish certain tasks is assisted
at times by the assistance icon that pops up offering suggestions
alternatives to accomplish the task.
Internet Agents
Internet agents that operate in the network environment can be used for
automating a variety of tasks that are associated with the internet. These tasks
may involve accessing, filtering, and even responding to information requests
on their own. Internet agents, based on the services offered, can be broadly
classified further as follows:
Web Search Agents: The web has emerged as vast resource of
information, with millions of pages of information online; it is not a
trivial task to locate a piece of information that may be of interest to the
user. These agents automate the task of accessing relevant information
for the user’s requirements, and then filter it out, based on the acquired
knowledge, through the past actions, behavior and profile of the users.
Web Server Agents: These agents reside in and assist web servers by
offering agent services. These agents include the interpretation of
requests by other agents, and responding according to the agent
interaction protocol, to facilitate agent based electronic commerce.
These agents reside at a specific web site, to provide agent services.
Information Filtering Agents: These agents are used for filtering out
electronic information, according to a user’s specified preferences. A
simple example of such an agent is an electronic mail filtering agent that
can be configured to sort the incoming mail into multiple folders, based
on various attributes and contents such as the subject, author, and
priority.
Information Retrieval Agents: These agents deliver a personalized
package of information to the desktop, according to user preferences.
Based on the user configured preference, these agents wander around the
internet to gather the information and then filter and customize it for
delivery. An example of this type of agent includes customized news
delivery agents that explore various news sources around the internet,
gathers the information, and construct an electronic newspaper for the
user, based upon his/her preferences.
Notification Agents: Internet agents that notify a user of events of a
personal interest to him. The user can enable these agents to keep track
of the changes in information. A user interested in monitoring changes
in a web site’s content can activate a notification agents; any time the
content of the web site changes, it notifies the user.
Intranet Agents
A variety of agent services can be used in the intranet environment as well.
These agents track resources, events, and information in the limited
environment. The work flow automation in an organization uses the
organization’s intranet to monitor, facilitate and keep track of the work flow.
The agent technology can be utilized to automate the tracking and filtering of
many of the routine work flow processes among business entities in an
organization. These agents can also be deployed for providing intelligent
guidance services to users of enterprise database resources. Resource
brokering is another area where agent services can be utilized for performing
optimal resource allocation, in client/server architectures.
AGENT TECHNOLOGIES
The agent, in an electronic commerce environment, has to operate in a vastly
unstructured, distributed, yet connected universe of the internet. Thus, agent
technology should be able to effectively deal a variety of issues emanating
from with a variety of platforms, syntax and semantics of agent interaction
languages, and cooperation and control mechanisms adopted by independent
agents trying to meet objectives. The heterogeneous structure and
uncontrolled topology of the cyberspace poses challenges for agents trying to
move around in the cyberspace. Some technologies that have tried addressing
the issue include Jini, Discovery and Trader Services, and XML
Metadirectories. The other important issue relates to interfaces and languages
for defining the rules for inter and intra-agent communication. The XML,
Knowledge Query Manipulation Language (KQML); shared semantic bases,
and Agent query interfaces, based on COM and JavaBeans, offer solutions to
various technological challenges. Most of the technologies supporting today’s
agent mediated electronic commerce systems stem from Artificial
Intelligence (AI) research.
To develop a better understanding of the technologies needed to support
agents, an overview of the agent’s computing environment is required. In the
agent computing environment, a user can store information and preferences
in a knowledge base. Domain specific knowledge will consist of general
guiding principles. The process by which an agent performs its duties is
determined by the preferences of the user and the model behavior, based on
the constraints in the computing environment. The agent must have a clear
knowledge of the environment in which it is operating, in order to put its
accumulated knowledge to use. The environment must allow agents to query
other agents performing similar tasks. In the rest of this section the agent
environment consisting of agent languages, protocols, inter-agent
communication, coordination, knowledge and reasoning, and control and
search techniques are discussed.
Agent Languages
Various languages have been developed for defining intelligent agents and
the processing required for operating these agents. A brief overview of some
of the languages is given here.
Knowledge Query and Manipulation Language (KQML)
KQML is a language and protocol for exchanging information and
knowledge. It is both a message format and a message handling protocol to
support run-time knowledge sharing among agents. It can be used as a
language for an application program to interact with an intelligent system or
for two or more intelligent systems to share knowledge in support of
cooperative problem solving.
KQML defines a set of messages, also referred to as performatives. These
performatives define the operations that agents are permitted to attempt on
each other’s knowledge and goal stores. Performatives define a low level
layer, that is used for implementing higher-level models of inter-agent
interaction, such as contract nets and negotiation.
Telescript
Telescript is a commercial product, developed by General Magic
Incorporated that supports mobile agents in an electronic marketplace. The
language is an object oriented programming language in which state oriented
migration is seen as the basic operation—provided by the ‘go’ instruction and
a ticket argument-that determines the destination site in “varying levels of
specification”. A Telescript engine exists at each site to accept and
authenticate migrating agents and to restart the execution of agents at the
statement, immediately after the go command.
The Telescript programming language lets developers of communicating
applications define the algorithms that agents follow, and the information that
agents carry as they travel the network.
Java
Java, developed by Sun Microsystems Incorporated, is an object oriented
language that is very reminiscent of C++. Java code is compiled to a
platform-independent byte code, for portability, but migration and dynamic
extensibility of the byte code are not explicitly supported. The object oriented
nature of the language makes it highly desirable, since a generic agent class
could be developed and other agent types (for example, a domain agent)
could be specializations upon that class, for example, the
JavaAgentTemplate.
Tool Command Language
TCL (pronounced ‘tickle’) was originally designed to perform the tasks of
traditional scripting languages; the creation of macros or code segments that
link compiled applications together. However, more recently, TCL has been
proposed as a language for writing mobile agents. Unfortunately, since TCL
is a scripting language, its inherent support for migration and dynamic
extensibility is non-existent. Also, since the language is interpreted directly
from source code it is also disadvantaged, due to the fact that it may not be
wise to allow other people to inspect the source code of agents. However,
despite these disadvantages, TCL is being used, and a proposal has been put
forward for a safe version of the language, called Safe TCL.
Agent Communications/ Requests
As agents are processes that operate in a distributed environment, commonly
used technologies that involve inter process communication are often used
for communication amongst agents. Broadly, these technologies can be
divided into three categories:
Synchronous Communication Oriented RPCs: Remote Procedure Calls
(RPC), which is a generalization of the traditional procedure call, can be used
for the request-response cycle of agents that communicate with each other, as
RPCs support communication among procedures that reside in different
locations. The RPC interfaces provided by most distributed operating systems
environments can interlink compiled procedures residing in different
machines. Even though compiled solutions are the most efficient and require
the least resources, they place severe constraints, as most decisions will have
to be taken while building the agent.
Asynchronous Message Oriented Techniques: Remote programming,
which is based on message passing techniques, can be used for process to
process communication in distributed environments, where agents pass
messages containing data and control to communicate with each other. This
model uses a loosely coupled approach, whereby, a single client to server call
is used to store and retrieve data, thus achieving a higher level of abstraction.
Database Middleware: This is a software layer that provides access to
homogenous and heterogeneous databases across multiple protocol
environments, and communication protocol conversion. It offers greater
flexibility as changes can be made to the system without the need for
recompilation, but the complexity arising out of adding an additional
component to the distributed system poses a constraint.
Agent Coordination
Agents operating in an open ended internet environment, trying to meet a
goal, may communicate with multiple numbers of other agents. They
consume a considerable amount of resources on the systems they run. An
agent may take disk resources to store information, use memory storage, and
a high level of computing power to perform the task at hand. It may not be
obvious to the system owner, as agents mostly run in background. Thus, it is
important to have some kind of coordination mechanism that will hinder the
unchecked growth of inter agent communication. The following two
approaches are predominantly used for coordination among multiple agents
in a distributed environment.
Contract Net Approach: In this approach, agents distribute requests for
proposals. Recipient agents process and evaluate proposals, and submit bids
to the originator. The originating agent processes all bids and evaluates bids
according to the rule base or predefined knowledge criteria, to come up with
a ranking of all the bids. The originating process then awards the contract to
successful agents.
Specification Sharing Approach: In this approach, agents publish
information about their capabilities as well as needs. The published
information is shared amongst other agents, and used for coordinating their
activities.
Agent Reasoning Capability
Typically, systems use a combination of statistical, machine learning, neural
networks and, inference techniques, to acquire reasoning capability. Any
agent system is implemented in stages. In the first stage, the system is trained
with rules or training data. The training is done by either by feeding the rules
or by providing a large set of example data with the right answers. The
training data is used for calibrating the reasoning ability of an agent. There
are several approaches that are used for building agent reasoning capability.
Rule Based Approach: In this technique, agents use stored rules to
determine the action they should initiate for a given situation. The rules may
describe the condition/situation and action. A simple example of this in use
can be seen in e-mail filters. The email filtering rules are of the form IF
{Conditions} then {take-this-action}. For example, “if the From: field of the
email has [IIML] Then move it to the IIML folder”. In a general system,
these rules, also called production rules, are made up of the two parts. The
left hand side describes the condition and the right hand side specifies the
action associated with the rule. An agent system usually has multiple rules. In
a multiple rule agent a situation being processed may trigger a rule, whose
action in turn may trigger another rule and so on. The chain effect of
triggering multiple rules is also called forward chaining in Artificial
Intelligence literature. In this type of system, users/trainers must recognize
where an agent would be useful, program the agent with rules based on the
set of preferences, and must also change these rules when the preferences
change.
Knowledge Based Approach: A knowledge engineer—expert in the
application domain—compiles a large amount of information in a specific
area. This knowledge base is then provided to the agent to deduce appropriate
behavior, depending upon the incoming situation. The method involves
substantial work on the part of the knowledge engineer, to endow the
program with a substantial knowledge in the given domain. Even with the
expertly endowed knowledge, in the beginning agents require constant
learning and updating of newer situations resulting from experience.
Simple Statistical Approach: In this method, agents learn from the
substantial statistical history. By analyzing the accumulated information,
statistically, agents determine the temporal as well as non-temporal
correlation among events. This information is used by agents for predicting
behavior in future events.
Neural Network Approach: The neural network approach mimics the
functioning of the human brain; it organizes the knowledge in a set of
interconnected nodes, and forms a web. The neural networks learn from
experience. Thus, they require training data and scenarios to compute the
weights of the nodes in a neural network. Typically, a neural network is
organized in three layers—an input layer, an output layer, and a hidden layer.
Each of these layers is made up of a several processing nodes (neurons).
These networks require a large amount of training data to develop the right
patterns, that can represent the non-linear mappings between input and the
output patterns.
Agent Control
The agents, in trying to perform the task, are driven by pre stored knowledge
and rules, and gather information in a distributed cooperative environment by
occasionally contacting or activating agents on other systems. Agents will
have to be controlled to ensure that they increase productivity and do not
create chaos. One method of ensuring control is by specifying the duration
for which an agent will perform a certain task. Another method is to allocate
resources to the agent, prior to its dispatch, to ensure that is does not use
resources disproportionately.
User Interfaces
Traditional shopping experiences vary, depending upon the needs of the
consumer and nature of the product offerings. Matching the system’s user
interface with the consumer’s manner of shopping will provide an easy to use
mechanism for the user to interact with an agent based or mediated shopping
system, resulting in greater customer satisfaction. The user interfaces that are
offered by most systems today are similar to those in online electronic
catalogues, but these do not offer a familiar ground for shoppers. Three
dimensional views of shopping malls, through use of VRML, can provide the
required familiarity, but are constrained by problems of navigation and
bandwidth.
AGENT APPLICATIONS
Agent based commerce is positioned at the highest level of user interaction,
because it utilizes all other levels of the web information hierarchy to
accomplish a specific task. As a starting point, a proposed segmentation of
these solutions point to four different approaches. Here are the four classes of
agent based commerce applications:
Automated-pull: These agents concentrate on assisting users in finding
precise information. The precise information is determined based on ad hoc
or pre-defined needs. In most of the cases, the agent utilizes the browser as
the interface for interaction. These agents carry out parallel pulling of the
information from the web resources and filtering them, based on specified
requirements and the pre defined user profile.
Web Automation: Web automation agents treat the information on the web
as an inventory of applications. These agents automate the process of
integrating a software application with the web, for a specific purpose, which
can then be replicated as and when desired. The aggregate application is built
using web automation tools.
Interactive Personalized Catalog: These agents integrate heterogeneous
sources of information from different information catalogs and present the
user with a real-time, personalized view of a new, integrated marketplace.
Information Filtering: These agents focus on personalizing user
preferences, based on a pre-determined profile that adheres to the Open
Profiling Standard (OPS), a new privacy standard. They are usually
integrated transparently within a web site.
Agents can be used in both Business-to-Consumer and Business-to-Business
transactions.
Agent Used in Buying and Selling on the Web
Today information about products and vendors is easily available on the web,
and orders and payments are automated, but there are several stages in the
buying process, such as information collection, buying decisions, purchase
and payment, where humans are involved. Intelligent software agents can be
used in certain stages of the process. This not only reduces the transaction
costs, but also improves the entire experience for the buyer.
Buying agents automatically collect information about vendors and
products that meet specific needs, evaluate the offerings, take decisions about
vendors and products to investigate, negotiate the terms of the transactions,
place orders, and make automated payments. The buying process of the
consumer consists of several stages. In a typical buying process, the first
stage consists of articulation of the consumer’s need. Consumers, in most of
situations, do not specify their needs explicity. The process of capturing the
intensions and/or preferences of user needs is of utmost importance to agent
functioning.
Agent technology has not made great progress in the need identification
stage and currently agents can only help in repetitive purchases. A
notification agent, called ‘Eyes’, at the Amazon.com site monitors a
catalogue of books and notifies customers when books of their interest are
available.
Once the consumer need has been identified, may be with the aid of a
monitoring agent, the process enters the product brokering stage. In this
stage, several agents carry out critical evaluations of the product information
and make recommendations to customers. Search techniques such as content
based filtering, constraints based filtering, and collaborative filtering can be
deployed. The more-difficult to characterize products, like web pages and
restaurants use collaborative filtering agents such as PersonaLogic and
Firefly. Apart form the above two techniques, simple rule based techniques
and data mining techniques are also used at this stage.
In the merchant brokering stage, the product brokering model compares,
evaluates and ranks product alternatives according to the consumer specified
need-based preference structure. Bargain Finder compares prices from
different merchant web sites and makes recommendations accordingly. As
comparisons are carried out only on the basis of price, a large number of
merchant’s block these types of requests.
Today several sites require their customers to manage their negotiation
strategies on their own, over extended periods of time. It is here that agents
play a vital role in automating the process of negotiation. AuctionBot,
Kasbah and Tete-a-tete are agent systems that help customers in negotiations.
Thus, first generation agent mediated e-commerce systems are creating new
markets and reducing business transactions costs.
The negotiation involves two or more parties that jointly search a space
for possible solutions, with the goal of reaching a solution, that satisfies all
the parties, or evolving a consensus. One important area of the transaction
that requires negotiations is the price and terms and condition of the
transaction. Stock markets, auctions, and flea markets (bazaars) are
transacting places where negotiation is used in traditional commerce. The
benefit of dynamically negotiating a price of a product, instead of fixing it, is
that it relieves the merchant from needing to determine the value of the goods
apriori. Rather, the burden of determining the price is pushed into the market
place itself. Consequently, the limited resources are allocated fairly i.e., to
those who value them the most. However, there are impediments to using
negotiation as a means of determining the value for commerce. For example,
in the physical world, certain types of transactions—like those in auction
houses—auctions require that all parties be geographically co-located. Also,
negotiating may be too complicated or frustrating for the average consumer.
Finally, in some protocols negotiations occur over an extended period of
time, which does not cater to impatient or time-constrained consumers. In
general, real world negotiations accrue transaction costs that may be too high
for either the consumers or the merchants.
Fortunately, many of these impediments disappear in the digital world.
For example, www.OnSale.com and www.eBay.com are two popular web
sites that sell refurbished and second hand products, using a choice of auction
protocols. Unlike auction houses, these sites do not require that participants
be geographically co-located. However, these sites still require that
consumers manage their own negotiation strategies over an extended period
of time. This is where agent technologies come in. In the following sections,
we introduce some of these agent technologies.
Kasbah
Kasbah, is an electronic marketplace, where agent programs carry out
transactions with each other, on behalf of the consumers. Kasbah implements
a consumer-to-consumer electronic commerce system, where agents buy and
sell. The transactions are based on continuous double auction mechanisms.
Whenever an item is to be sold in the Kasbah, a new agent is created. The
new agent is provided with the description of the item to be sold. In addition
to this, a set of parameters, used to control the behavior of the agent, are also
specified. These parameters include:
Desired Date to Sell the Item: People usually have a deadline by which
they want to sell the item.
Desired Price: The price at which the consumer would like to sell the
item.
Lowest Acceptable Price: The minimum price at which the consumer
will sell the item.
These parameters define the agent’s goal. Armed with the desired price as
well as lowest acceptable price the agents works on achieving the goal of
fetching the maximum possible price in the given time frame in which to sell
the item. The process and mechanism through which the agent achieves the
goal is determined by the agent itself.
These agents are proactive, and once launched they try to sell the goods in
the market place, by contacting other buying agents and negotiating the best
deal with them. Selling agents start negotiations at the desired price, keep
lowering their prices, and on the due date, they try selling it at the lowest
price. It is possible that there will be no buyers, in which case the agent fails
to achieve its goal. The consumer can then check on his/her selling agents
and determine which other agents the selling agents had made contacts with,
and what prices have been offered by these agents. This information might
prompt the seller of the item to lower an agent’s price parameters, if they see
that the offers coming in are much lower than expected. The consumer/owner
of an item always has the final control over his agent.
BargainFinder
BargainFinder is an experimental virtual shopping agent for the web,
developed by Andersen Consulting. BargainFinder uses parallel query
architecture, similar to Meta Search engines, and heuristic topic phrase
extraction techniques, to query the price and availability of user specified
music CDs. The comparison shopping agent takes the user’s product query
and submits it in parallel to a group of online vendor’s, by filling out the form
at each site. The agent collects the query results, parses the results and filters
out the header, trailer, and advertisements information, to find each vendor’s
price for the product. The agent then collates the filtered results and presents
them to user in a summarized form.
Comparison shopping agents extend the reach and price discovery
capacity of shoppers, and offer the following advantages.
Each vendor may organize the information on the internet based shop
and shopping catalogues in their own unique way. The agent extracts the
relevant information from these vendor sites, sparing the user from
navigating different vendor sites and dealing with separate user
interfaces, to extract the relevant price and availability information.
In the price discovery phase shoppers have to search for information by
visiting multiple vendor sites, extracting, and comparing the price and
availability information on their own. The agent can automate the task
of locating the relevant vendors, extracting, and ranking the price and
availability information. The shopper can provide the product
specifications to the agent. The agent, working in the background, can
collate the information and present the summarized result, thus relieving
the user.
Comparison shopping agents can work better if the information from
vendors can be readily extracted, but there are certain hurdles that these
agents need to clear:
In a competitive marketplace, many virtual stores do not want the
shopping to be based purely on the price and availability information.
As a result, they are reluctant to allow agents to extract the information
for shopping purposes.
Also, as the agents themselves are evolving, agent users may not be
willing to fully rely the on the agent’s ability to notice sales and special
promotions. For instance, a software product that interests a user may be
part of a software bundle, for a slightly higher price, and may be missed
by a shopping agent.
There have been two approaches to address the hurdles described above.
These approaches are based on vendor cooperation and machine learning
even when there is no cooperation from the vendors.
Cooperative Agent/Vendor Model: In this approach, there is tacit
cooperation between agents and vendors and the vendors put the product
information, that they would like to share with agents, in a standard form.
The Identify Markup Language (IDML) extension of the HTML is one such
an attempt. The IDML offers the capability to marketers and vendors to
specify how they want to be identified and how they would like their brands
and products to appear to searchers. This gives vendors control and ability to
specify what products can be directly accessed. In essence, IDML gives
vendors a structured way to identify their products. Agents can make use of
this structured information for collating the needed information.
Machine Learning Approach: In the machine learning approach, the agent
parses and learns the structure and content by parsing the information
available at vendor sites. This approach is implemented in ShopBot. The
ShopBot agent attempts to learn how to shop at virtual stores without any
cooperation from vendors. ShopBot uses the machine learning approach to
find the HTML forms with product information, at a vendor’s site. The agent
uses the information available on HTML forms to identify the product
information that matches with the user query.
Agent based shopping is still evolving, in the long term better and robust
solutions will emerge as retailing on the web goes beyond today’s functional
and replicates, by and large, mail order catalogs.
An example of the collaboration based electronic commerce agent,
FireFly, is described here.
FireFly
FireFly is a collaboration agent that makes recommendations, based on the
group evaluation of products. Collaboration agents induce users to explicitly
evaluate specific products. The agent compiles the evaluation as well as
profile of the evaluating users. Its product recommendations take into account
the compiled information; thus, it recommends a product based on the
preferences of people with similar profiles. To users looking to buy a
product, the Firefly agent also recommends other products that have been
bought by those who purchased the product being assessed by the user. In
essence, collaboration agents try to capture “word-of-mouth” advertising.
FireFly uses memory based reasoning to find user pattern clusters.
Memory based reasoning operates on situation-action pairs. For instance,
while shopping for music, the situation would be described by the artist,
album, and associated attributes such as the genre of music, whereas the
action would represent the user’s like or dislike of the album in question.
Memory based reasoning, based on the nearest neighbor algorithm, enables
recommendation. The algorithm uses a distance metric, that computes the
weighted sum of the distance between the corresponding attributes of two
situations, to determine the similarity. Memory based reasoning is in essence
a case based reasoning technique, where every user action is used as a case
entry.
There are other methods for building collaboration agents. For instance, a
domain expert can use a rule based system to encode recommendations, after
mining user data offline, using clustering techniques. Another approach
supports a distributed registry of user interests, while preserving privacy. In
such an environment each user agent keeps track of other user agents it
encounters, and bootstraps itself by asking for referrals to other agents in
order to find other users that may match the specific interest of a given user.
The advantage of this agent, lies in its scalability, as it does not require a
central registry of user interests.
AuctionBot
This is a general purpose internet auction server at the University of
Michigan. It supports multiple auction types. In AuctionBot, a seller can
create new auctions to sell products by choosing from a selection of auction
types, and then specifying its parameters such as clearing times, method for
resolving bidding ties, and the number of sellers permitted. Bidders can then
bid according to the multilateral distributive negotiation protocols of the
auction. In a typical scenario, a seller would state the reservation price, after
creating the auction, and let AuctionBot manage and enforce buyer bidding
according to the auction protocol and parameters. What makes AuctionBot
different from most other auction sites, however, is that it provides an
application programmable interface (API) for users to create their own
software agents, to autonomously compete in the AuctionBot marketplace.
Such an API provides a semantically sound interface with the marketplace.
However, it is left to the users to encode their own bidding strategies.
Tete-a-Tete
This is another agent that uses the negotiation approach to retail sales. The
Tete-a-Tete agent, instead of following the competitive negotiation strategy,
uses a different approach, based on cooperation. The cooperative approach
permits Tete-a-Tete agents to negotiate multiple transaction terms such as
warranties, delivery times, service contracts, return policies, loan options, gift
services, and other merchant value added services. The Tete-a-Tete agents
follow the argumentative style of negotiation protocol in a bilateral and
multi-agent negotiation environment.
Notification Agent
A notification agent informs/notifies users of significant events. The
significant events, usually specified by users, can be made up of a change in
the state of information such as:
content change in a particular web page.
search engine additions for specified keyword queries,
user specified reminders for personal events such as birthdays.
Internet notification agents can be server based programs that keep
monitoring user specified sites, or can also be desktop based programs that
try to provide the same functionality. Examples of desktop notifier programs
include NetBuddy and SmartBookmarks. These notifier programs essentially
monitor the specified resources and alert the user as and when any change
takes place in these monitored resources. Desktop based agents suffer from
two major disadvantages- the first is the computational burden they place on
client CPUs and the other is the inefficient use of the limited bandwidth. On
the other hand, server based notifiers make better use of bandwidth by
combining the interest of multiple users of, many of them trying to monitor
the same resources. Irrespective of desktop or server based agents,
notification agents offer a great deal of efficiency to users trying to keep track
of these resources manually, by increasing user productivity and reducing the
number of HTTP connections—since desktop notifiers need not fetch the
entire document.
Notification agents monitor change in information by employing one of
the following methods:
HTTP ‘if-modified-since’ Request This is a special Header Request that
returns a document only if the page has been modified since the
specified date. This is a fairly inexpensive operation involving one
HTTP connection and a couple of hundred bytes of information transfer.
Text Only Retrieval As some of the change in information such as
advertisements, dates, counters, etc. is not of interest to a user,
notification agents retrieve only the text of a page, without the graphics
and hyperlinks, and parse the retrieved text to determine any change in
the published information.
Embedded HTML Extensions These are directions to notification agents
embedded in HTML documents, from publishers. These can be placed
in ‘head’ protocol fields as ‘meta’ tags, the document heading, or in the
body of the document. For instance, an extension tag may instruct the
notification agent to ignore any change in a particular document section.
Embedded HTML extensions require the cooperation of web publishers.
Although this may seem an additional burden to webmasters, such a solution
is a good model for businesses selling a large number of products through the
web. Although HTML supports meta tags, it does not introduce any standard
for document or product attributes. One such attempt to address this issue is
the IDML extension to HTML. IDML is a set of HTML extensions that lets
publishers specify who they are, what the web site is about, and the products
for sale, using a standard format.
URL Minder
This notification agent retrieves web resource periodically to detect changes
from the last retrieval. A user registers a web site of interest using a form.
The URL minder monitors the specified web resource, and sends the user an
e-mail message whenever it detects a change. It thus relieves the user from
having to visit the site regularly to check for changes. The URL minder can
also keep track of the search results of a certain query; if and when the query
results from search engine changes, it can inform the user. It also monitors
each registered page once a week for changes. It checks for each retrieved
documents, using the Cyclic Redundancy Check algorithm. In order to filter
the imprint of cosmetic changes on the computed signature, the URL minder
has HTML extension tags instructing its robot to exclude specific sections of
a document.
Mobile Agent
Concordia
Concordia is a Java based framework for implementing mobile agents. It
requires that a separate, lightweight Concordia server runs on each
participating machine, on the network. In this environment, mobile agents
migrate from one server to another, in order to perform the task. Mobile
agents travel to other servers using the Java object serialization mechanism
for transferring data. The serialized mobile agent is converted back into Java
objects, which are recreated in the new host’s Java Virtual Machine.
The agent carries with it a program itinerary that consists of a list of
destinations and a set of actions to perform at each destination. The
destinations specify Concordia servers, while the set of actions specify the
methods to be run at the remote Concordia server location. In most mobile
agent frameworks, the agent has a ‘go’ method that executes when it arrives
at each server. In Concordia, however, agents can execute any method, on the
destination server, that is available on the remote server. It also allows for the
dynamic generation of itinerary at run time. The Concordia framework
supports the run time dynamism where the itinerary itself can adapt and
evolve, depending upon the tasks performed and the data that is gathered.
Thus, an agent can modify its behavior according to the outcome of events,
during the course of its journey, endowing it with more intelligent behavior.
For example, a data gathering agent could find the right expert at a
professional services company on a four server itinerary, that spans the
geographic boundaries of the organization. The itinerary could dictate that
upon completion of an agent goal, the agent should return the information to
the user.
Concordia mobile agent framework is designed to support the robustness
and reliability needed for enterprise solution applications. It is geared for
providing mission critical security and reliability features required for the
large scale applications. The Concordia security model combines symmetric
and public-private key encryption, used to protect agents during network
transfers. Every agent in the Concordia framework represents a user and
hence is authenticated. Each agent carries the credentials of the user it
represents, in the form of a X.509 digital certificate. Permissions and
authorizations are granted based on the authenticated digital certificate.
Hence, an agent can perform only those actions that the authenticated user is
allowed to perform. It also implements a transparent mechanism for ensuring
reliability in the framework, thus, the application code is shielded from
failures at the server and/or network level. In case of agent failures, a
checkpoint-restore mechanism is used to restart agents. The recovery
mechanism relies on the state of the check point information. The before and
after execution state information is stored for each agent on a server. Anytime
a server is restarted, the recovery process is executed, which in turn restarts
any agents that had unfinished work left on the server at the last shutdown or
failure.
Agent in Supply Chain Management
Business-to-Business transactions have been the fastest growing segment of
electronic commerce. The growth and adoption of the world wide web based
transaction mechanism, by large corporations like General Electric (GE),
Cisco, and Intel, has already proven the benefits accruing from it. The initial
effort of General Electric (GE) to streamline the supply chain process,
through the deployment of electronic commerce, has already given way to an
electronic market place with thousands of suppliers conducting billions of
dollars worth of business, on what is now referred to as the Trading Process
Network (www.tpn.com). As the level of integration of activities between the
supply chain partners continues to increase, the number of suppliers and
supply chain webs may grow, and it will not only be complex but difficult as
well to get the best available and negotiated deal manually. This new
complexity has already began to exert pressure on devising automated ways
to scan the market space, generate options, negotiations, and support decision
making.
Agent based systems have already proven their utility in consumer market
places, as seen in preceding discussion. Agent based systems have also been
effective in adapting to dealing with the planning and scheduling of the
manufacturing processes. A combination of these two technologies can be
used for building viable agent based systems, that can be of assistance in
automating the processes of sales, procurement, collaborative forecasting,
design, and planning.
FUTURE
Agents are small pieces of software code that can automate many a tasks. On
the electronic commerce front, they can be used in web sites to enhance sales
and customer support, by customizing pages for individuals. Agents can also
assist in promoting sales, by capturing customer preferences and guiding
them to suitable products. In case of customer service, agents, based on the
problem description and customer profiling, can route help desk requests to
the right places. Whatever the purpose, it is necessary to build many agents,
with different kinds of intelligence, to handle the situation.
As intelligent agents in the shopping arena are likely to contact many
traders and suppliers, based on exchanged information, the concept of store
loyalty may suffer, as they are likely to place more trust in brands while
optimizing and negotiating for price information. The store is likely to play a
secondary role and may see erosion of some of its identity, because in agent
based commerce the physical environment which a consumer enters and
spends time in, turns irrelevant. Agents work towards the goal of matching
user preferences with product availability, and summarize the findings for the
user, in a ranked order. The marketer and selling agent thus need to be trained
to find a way to sway consumer’s preferences in their favor. Consumer
automated agents may end up ignoring many of the storefronts from where
they are unable to extract the price availability and terms of transaction
information, in addition to the other product attributes.
As we advance toward agent oriented commerce, and agent based
societies, the product seller and buyers are likely to get closer to the true
market value of products, in real time. It would be possible to better quantify
the effects of advertising and marketing promotions, and the effects can be
observed quickly. This implies that product suppliers also have to react to
changes in an accelerated manner, inducing a company to continually change
and update their products and operations.
Although, the concept of the intelligent agent has been around for many
years, actual implementation is still in a very early stage. In practice, we are
able to construct agents with relatively simple intelligence. As agent
technology finds wider adoption and acceptance, agents may evolve to
contain complex reasoning, and may become very sophisticated. These
sophisticated agents hold, potential to reduce “surfing” to a great extent, as
they will automate information gathering, option generation, negotiations,
and purchasing decisions, for both buyers and sellers. The users—rather than
scan and surf enormous number of sites, to locate the best deal—will be in
position to ask their agent to start searching for the best deal for a given
specification. The agent can collect and rank the information in the back
ground, and present it to the user when be come back to find the information.
Ultimately, consumers will have their own personally trained shopper and
research assistant, who knows all preferences, goals, and information desires.
The technologies and approaches highlighted here serve to show only a
few examples of how agents are definitely metamorphosing the way we
interact with the web. The long term role of agents in electronic commerce
will be transformational, akin to that of search engines on information
discovery, over the world wide web.
The amalgamation of the capabilities of agent-based technologies to that
of information appliances, beyond browsers and existing applications, will
have a deep impact on commerce. Information appliances with embedded
agent capability can become highly specialized point-of-sale devices for a
variety of products and services as they can scan market space buyers and
likely sales by interacting with market information databases and other buyer
agents. For example, a telephone device with agent capability can
automatically scan for changes in addresses and phone numbers of friends
stored in it, and keep them updated in the personal online directory. Or, a
simple personal digital assistant with wireless connectivity and agent
capability can scan information databases as per user preferences, and keep
an up to date status of things that user may interested in. For a user interested
in entertainment, it can maintain an up to date status of movie listings in the
neighbourhood cinema halls, restaurant reviews and deals and promotions
running there, special games scheduled, and plays running in theaters near his
location.
Software agent technology has the capability to affect people’s life
greatly. Agent technology will not only alter the way in which we interact
with computers but also the way in which we conceptualize and build large
systems.
SUMMARY
Intelligent agents can increase user productivity by carrying out certain
programmable routine tasks in the background. Electronic commerce is
creating such a vast market place, with enormous numbers of products and
pricing options. As a result, a product search and price discovery that meets
the satisfaction level of the user are becoming increasingly time-consuming
processes. In this chapter, we describe agent technology and type of functions
they can automate. Agents can be classified based on several attributes such
as mobility, intelligence and autonomy. This chapter describes agent
technology and the standards and languages used for defining and operating
agents. Finally, various agents that have been prototyped, implemented, or
have evolute into product offerings are described in the chapter.
REVIEW QUESTIONS
1. What is an agent and how can it be used in the electronic commerce
environment?
2. What is meant by autonomy, in the context of agent definition?
3. What are mobile agents? How do they differ from static agents?
4. Describe the key characteristics of agents.
5. What are internet agents? Provide a few applications of internet agents?
6. Describe role of selling/shopping agents in electronic commerce. Give
two examples of the selling agents.
7. What are collaboration agents?
8. Illustrate with an example, the purpose and functioning of a negotiation
agent.
The Enthusiasts come into the ‘Enthu’ zone. The Waverers fall into the
‘Neither here not there’ zone. The Mules and Terrestrials fall into the
‘Inertia’ zone.
In the case of the two most commonly used business models, portals and
e-retailing, portals operate in the ‘Enthu’ zone and e-retailers operate in the
‘Inertia’ zone. As Fabmart focused on e-retailing, it operated in the ‘Inertia’
zone, hence attracting traffic and increasing the number of orders was a great
challenge. In this model, revenues were through margins obtained from
selling items online.
Vision
Even though the team comprised of experienced professionals from the
Information Technology and the Internet industry in India, they were very
clear that the focus was on retailing, and not on internet technology. There
was a conscious decision to exclude the ‘e’s and ‘.com’s. The competition
was not only online retailers but also brick and mortar stores involved in
retailing similar categories. The objective was, thus, to get a part of the bigger
pie.
The vision was to be India’s finest online retail brand. It was decided that
Fabmart would use the leverage offered to the advantages of the internet to
offer consumers a great shopping experience.
In the three year business plan of Fabmart it was stated that the target for
the first year of operations would be 10,000 happy customers. Fabmart
projected a sales target of ` 45 crore by the third year, at it, which intended
point to break even.
Strategy
The goal was to build a virtual supermarket. The objective was to:
Provide a great shopping experience to customers, so that repeat
purchase is possible
Gain a remarkable first mover advantage
Get associated with online shopping and create strong entry barriers
There were two approaches to achieve this. In the first approach, the
various categories of items to be sold online would be launched together. The
store would have to launch books, CDs, toys, garments, and items in other
categories simultaneously. It was felt that this would lead to an average
collection of items in each category. Customers would visit the store for the
first time, but would not come back again due to an average collection, as
they does not gain substantially by shopping online. The second approach
was based on phased growth in which as a first step, the online store was
built for a specific category. After the completion of the first stage, a new
category would be launched as a second step, and then the thired category,
and so on. This would provide an opportunity to build a good collection of
items in each category, leading to greater customer satisfaction and repeat
purchase.
Vaitheeswaran stated the objective as:
“Each store on its own must make sense to the customer. Customers must
be able to recall Fabmart among the top 3 stores in each category.”
A total of 21 categories were listed for this purpose and it was decided to
have all the 21 categories up and running in 15 months. The 21 categories
chosen fell under three major areas. The areas are:
Amenable to selling on the web, but are impulse based like music and
books
Need based like provisions
Impossible like cars and jewelry
The criteria for choosing new categories were as follows:
The industry type— growing or mature
Size of the industry
Internet friendliness of the product
Distribution mechanism
Tangible difference in comparison to physical buying
A total of six categories—books, music, gift, garments, provisions, and
jewelry–were identified and these covered the three areas under
consideration.
The following set of characteristics were used in selecting the first
category to be launched online:
Cost of trial for the customer must not be very high
Back–end logistics must not be very complicated
The store must not be able to tamper with the product
Internet technology must provide some advantage in terms of selling the
product (In case of music, the customer can listen to the music before he
buys cassettes or CDs)
No other store must have done a good job in the category before
Using the above parameters, two categories were chosen—books and
music. Rediff.com had already entered into the scene with books and hence
music was chosen as the category to be launched first.
In tune with the concept of virtual organizations, most operations at
Fabmart were outsourced. A set of four operations that were identified as
critical to the company is as follows:
Brand building
Relationship with music and book companies
Order fulfillment
Design of the store
The difference in the shopping experience between an online store and a
brick and mortar store is highlighted in Exhibit 4. Most e–retailers face
problems in the delivery process due to poor back–end logistic support. Also,
most. com companies focus on the first three phases shown in the exhibit.
Such stores would be able to attract customers for the first time, but in
attracting repeat customers, the delivery process will have to be as promised.
Even if the entire front end is perfect, problems in the backed logistics will
lead to customer dissatisfaction.
In the internet start–up business, speed is an important criterion in
determining success. Pioneers with innovative ideas have a clear advantage in
establishing themselves with customers. As several operations have to be
carried out simultaneously, large amount of funds have to be spent. The
Fabmart team made a conscious decision to adopt a leadership position in the
‘Spending Graph’, shown in Exhibit 5. This ensured that Fabmart established
itself before its competitors. This also increased the risk profile.
In deciding on strategic partners, their commitment to the business was of
prime importance. Partners were chosen so that there were no conflicts in
business and commercial issues. Pentagon is their advertising partner,
Mindtree worked upon the store front–end for the book and garment store
and Integra Tech Soft developed the music store.
Branding and Marketing
The team contacted potential customers, drawn from friends and associates,
to gauge their reactions to the business plan. Even though they were
confident that e–retailing provided a great opportunity, it was found that
security was a major concern in the minds of the customers. There were
several problems like inertia in buying, security concerns, credibility of
merchants, negative international press, and negative word of mouth related
to online shopping, and these had to be effectively countered.
The base of internet users was assumed to be 15 lakh, out of which about
10 lakh lived in the seven cites of Mumbai, Delhi, Chennai, Calcutta,
Bangalore, Hyderabad, and Pune. The entire marketing and communication
plan was aimed at the Waverers and the Enthusiasts. The model adopted by
Fabmart is given in Exhibit 6.
In an industry where portals and e–retailers were the two dominant
players, the factors that had to be considered in deciding the promotion
programs are listed below:
Portals need to spend less money as compared to e-retailers, for the
same traffic
Ad spend is determined by the share of voice
Portal’s spending was enormous
This led to the conclusion that a lot of funds were required to build the
Fabmart brand.
The basic objectives of marketing were:
To build a Fabmart brand with a positive image
Drive traffic, registration, and orders
Various promotion programs were carried out for brand building. Fabmart
created a record of sorts with the launch of a music album exclusively on the
internet. “Sarvasri”, a music album by acclaimed Carnatic vocalist Dr. M.
Balamurali Krishna, was launched online and was not available in other
music stores for about 10 days. This promotion program started driving
traffic to the site. A festival of rock music was organized, wherein the top 25
rock artists participated, their biographies were made available on the site.
More over, their albums were available at discounted prices, and customers
could also win rock music CDs. A carnatic music festival, shown in Exhibit
7, was also launched on the site after the rock festival. The basic aim of such
festivals was to attract different categories of music lovers to the site. The
biggest success came from the ‘2 for 2 promo’ where customers could get 2
cassettes for ` 2.
The PR agency Corporate Voice partnered with Fabmart in building a
positive brand image. The media used included print media like dailies, and
magazines, apart from hoarding and bus shelters in a few cities. Banner ads
were also placed in higher traffic sites like ‘Hotmail’. Advertising on TV may
be carried out at a later stage. A free downloadable MP3 “Cyber Viber” by
Remo Fernandes also helped in attracting traffic. By the end of
February1999, Amit Heri’s first chargeable downloadable MP3 as available.
The promotion attracted 3000 new customers on the site, with the number of
registrations increasing to 500/day and orders increasing to 300/day.
Logistics
The store has a single consolidation point at Bangalore. Fabmart has a tie-up
with, Blue Dart, the logistics provider for the delivery of cassettes/CDs to
850 cities across the country.
The steps in the order management process and given below:
Customer places order
Fabmart authorizes payment through Citibank for credit card orders, or
waits for Citibank to authorize the Citibank debit card orders
Fabmart places order in the music distributor
Distributor delivers items to the consolidation point
Consolidation point packages each individual order
Courier company picks up the packages from the consolidation point for
delivery
When categories such as books and gifts were added, more consolidation
points were required. Talks are on with Fedex for overseas shipping.
Consolidation points would be opened in Singapore, Dubai, and Silicon
Valley with outsourced agents in another two months. Also, with more stores,
in the future providing customization would be important.
In the music industry, logistics problems exist as most distributors do not
have automated operations. This problem is not present in the case of books.
Hence, tie-ups are required with companies/distributors who are well
organized and can provide online stock status.
Systems Architecture
The main server is hosted in Bangalore through Bharti BT. A Compaq server
with dual CPU is used for the web site with RAID level 5 built–in for
redundancy. The staging server from Wipro has 1000 audio clips are placed
on it and the SQL runs on the Compaq server. At the office, 64 Kbps leased
lines are used and for backup, an ISDN connection and dialup lines are
available.
The software behind the entire operation is the web enabled ‘Virtual
Inventory Model’(VIM). All authenticated orders are picked up by VIM.
Consolidation of orders is through the listing of all albums against the
specific company and printing out one single purchase order. This purchase
order is sent to the music distributor, Raaga, who delivers all albums to an
outsourced consolidation point at the end of the day. At the consolidation
point, the VIM software updates all orders that have been authorized and can
be serviced. The software also prevents shipping out of partial orders. In case
the music distributor is unable to provide a specific album, it is obtained from
other sources. The package shipped to the customer contains the cassette/CD,
invoice, a Fabmart pouch, and stickers. A pending order report is generated
for items that are not available at the end of the day.
Store Front-end
The Fabmart music store expects to have two types of customers:
The customer who knows exactly what he wants
The customer who is not sure of what he is looking for
To cater to both these segments, the store–front has a “Search” facility,
apart from providing listings of the various music categories. It has two
doors–Indian and International–as shown in Exhibit 8. Inlay cards containing
information about the composer and album, which is sealed in case of
conventional music stores, is available to the customer in the online store, as
shown in Exhibit 9. The ordering process is very simple. Long registration
forms have been avoided. Only information that is fundamental to the
business like name, address, and e–mail are mandatory. All other details are
optional. The aim is to help the customer order what he wants and also help
him recollect his albums of interest. A shopping cart is available for the
customer to put in all the albums. Also, a wish list, consisting of the
customer’s albums of interest is maintained at the store for 90 days. This
creates a big differentiation from the physical store. A customer can
recommend an album to a friend by providing this e–mail id whereupon
Fabmart sends a mail recommending the specific album. When an order is
placed, an order number is generated, which is used as reference for future
correspondence.
By registering, the customer opens an account using an userid and
password. Using this id and the ‘My account’ feature, the customer can track
the status of his order. The various stages, an order can be in are:
Pending for Authorization—This is the status till the bank authorizes the
payment
Pending for Allocation—This is the status after authorization and before
the album has been sourced
Ready for Shipment—This is the status after sourcing and before the
courier picks it up
Shipped—This is the status when the courier has picked up the package
Delivered—This is the status after the proof of delivery is received from
the courier
Order tracking is another area of differentiation from other online stores,
and provides value to the customer. Once the customer places an order, an e-
mail message with details of the order number, as shown in Exhibit 10, is
sent for confirmation. After the item is shipped, a second e-mail is sent with
details of the consignment number, as shown in Exhibit 11.
Security
Payments are through credit cards or Citibank Suvidha debit cards. An SSL
link, with 40–bit encryption, is used. After the tie-up with Citibank, the
Citibank Suvidha debit card could also be used. This opens the Citibank site
automatically and the transaction is over a SSL link with 128-bit encryption.
In this case, the PIN is given to the bank directly and the bank authorizes the
payment. As this system addresses security concerns, the major cause of
concern in India, an ad, as shown in Exhibit 12, was released in all dailies
and magazines. This resulted in good positioning as it addressed security
concerns, and also helped build the Fabmart brand. The 7-day return scheme,
in which customers can refunds get on goods that they bought from the store,
further helped in this direction. In this scheme customers will have to return
goods, with or without the reasons for the return, within 7 days of receipt.
Strategic Priorities
At present, the number of visitors has increased from 50 to 5000. There are
about 1 million hits per day with more then 300 orders being placed per day.
Initially, a six member team worked on the assignment. By March 1999, the
number of employees was increased to 20. People were required for running
different store categories. Books were the next area of focus and the book
store opened by the end of February, 1999. By March, the jewelry store was
up; the gifts/garments store was ready by April and the provision store by
May. With more stores in different categories being added, the backend
logistics is expected to get more complex. With traffic increasing, backend
systems and logistics are key areas of concern.
Fabmart would now require more funds, to the tune of 10–15 crore.
Financing, at this stage, should not be on the basis of cost of funds alone but
also on the strategic value brought in. An Initial Public Offering (IPO) also
needs to be considered. Thus, the success of the virtual supermarket would
depend on a few key decisions given above.
EXHIBIT 1
E–Business Space
EXHIBIT 2
Players in the E-Commerce Industry
EXHIBIT 3
Selling in the E–Business Space
EXHIBIT 4
Shopping Experience
EXHIBIT 5
Spending Graph
EXHIBIT 6
Marketing Model at Fabmart
EXHIBIT 7
Carnatic Music Festival
EXHIBIT 8
Store Front-end
EXHIBIT 9
Albm Details
EXHIBIT 10
E–mail message for Order Confirmation
Dear ………………,
We thank you for your order placed on the Fabmart store on Thursday,
January 06, 2000. The reference number of your order is 326323. You can
track the status of your pending orders at the store by clicking the ‘Track
Your Pending Orders’ (http://www.fabmart.com/music/receipts.asp) link on
your ‘My Account’ page.
We should be shipping the items ordered by you soon. We will send you a
confirmation of your shipment, along with the shipment details, as soon as
the ordered items leave our warehouse.
We thank you for shopping at http://www.fabmart.com. We hope to have
you visit us again.
Warm regards,
V. S. Ramesh
Vice President – Fulfillment
Shop at: http://www.fabmart.com
Browse. Shop. Have a great time.
EXHIBIT 11
E–mail after shipment of order
Dear ………………,
Your order, # 326323, has been shipped out from the Fabmart store on
Friday, January 07, 2000. The shipment will be delivered at your doorstep by
our fulfillment partner, Blue Dart Express Limited. The consignment number
of your shipment is D426332874. You can use this number to make enquiries
about your shipment with Blue Dart Express at their local office.
Please feel free to mail us at [email protected] for your
fulfillment/delivery related queries.
We thank you for shopping at http://www.fabmart.com. We hope to have
you visit us again.
Warm regards,
V. S. Ramesh
Vice President – Fulfillment
Shop at: http://www.fabmart.com
Browse. Shop. Have a great time.
EXHIBIT 12
Ad Copy to Address the Security Concern
__________________
Kavitha Rao, R. Srinivasan, and B. Bhasker prepared this case as a basis for
class discussion rather than to illustrate either the effective or ineffective
handling of an administrative situation.
1 Source: Business Today, “e–India’s e–biz models”, October, 7–21, 1999.
10Base2 113
10Base5 112, 113
10BaseF 114
10BaseT 114
A
Access Control 150, 217, 220, 224
Acknowledgement 125, 218
Acquirer 319–324
Active Server Pages (ASP), 184–186
Address Resolution Protocol (ARP) 123, 210, 211
ADO 185
Address Resolution Protocol (ARP) 123
Advance Research Project Agency 89
Advertising Model 50
Affiliate Marketing 372
Affiliate Model 52
Agents
Applications 475
Autonomy 471
Characteristics 471
Control 477
Cooperation 476
Coordination 476
Intelligence 470
Mobility 470
Agent Communications 475
Asynchronous Message Passing 475
Database Middleware 475
Remote Procedure Calls (RPC) 475
Agent Coordination 475
Contract Net 476
Specification Sharing 476
Agent Interface 472
Agent Languages 474
JAVA 474
KQML 473
Telescript 474
Tool Command Language 475
Agent Reasoning 476
Knowledge Based 476
Neural Network 477
Rule Based 476
Statistical Approach 476
Agent Standards and Protocols 477
Agent Transfer Protocol (ATP) 478
SATP 477
ALOHA 111
ANSI X12 71–73
Application Layer 118–119, 126
Application Level Firewall 219
ARP 123
Prevention of Spoofing 212
Spoofing 211
ARP Spoofing 211–212
ARPANET 89, 117, 118
Auctions 34, 57
AuctionBot 483
Authentication 93, 96, 97, 214, 237, 250–257
Authorization 238
B
B2B 16, 19
B2C 20–21
B2E 25
Banking 35, 36
Banner Advertisements 404
Customized 410
Effectiveness 408
Placement 405–408
Payment Model 405
BargainFinder 481
BITNET 89, 118
Blog Marketing 373
Brokerage Model 59
Buffer Stock 292
Business Case 302
Business Models 45, 46, 47
Advertising 50
Affiliate 52
Brokerage 59
Content based 47, 49
Definition 46
Digital Products 53
Electronic Store 58, 59
Freeware 49, 50
Infomediary 51
Information Content 48
Information Exchange 49
Internet Access 54
Manufacturer 60
Metamediary 56
Metered Service 55
Native 48, 53
Subscription 50
Transaction based 47
Transplanted 58
Web Hosting Internet & Services 55
Bullwhip Effect 360, 361
Business Service Infrastructure 70
Business-to-Business 16–19
Business-to-Consumer 20–23
Business-to-Employee 25
Business-to-Government 16
Buying and Selling Agents 479
C
C2B 22
C2C 22–23
CAT-3 Cable 106, 107
CAT-5 Cable 106, 107
Certificate 257–258
Certificate Repository 257
Certificate Revocation List (CRL) 257
Certification Authority 257
cHTML 325–326, 452–454
Ciphertext 325–326, 336, 365–367
Coaxial Cable 106
Cold Fusion Markup Language (CFML) 181
Color Map 190, 191
Common Gateway Interface (CGI) 165–169, 176–180
Alternatives 181
Security 229–233
Compression 190–193
Concordia 485
Confidentiality 238, 266–267
Consumer-to-Business 22
Consumer-to-Consumer 22
Content-Length 147, 148
Content-Type 140, 147, 148, 175, 180
Controller of Certification Authorities (CCA) 97–98
Coordination 295, 297
Coordination Cost 7, 8
Corporate Web-Sites 415
Cost Minimization 295
Cost-per-Action (CPA) 367
Cost-per-Click (CPC) 367
Cost Per Thousand Impressions (CPM) 400, 404
Cryptanalysis 239
Cryptographic Algorithms 243
DES 243–246
IDEA 245
RSA 246–248
SHA 249
Triple DES 245–246
Cryptography 239
Cryptology 239
CSMA 112
CSMA/CD 112, 115, 116
CSNET 89
Customer Service 282
CyberCash 228, 321, 330–336
CyberCoin 318
D
Data Encryption Standard (DES) 244–246
Data Integrity 253, 263, 266–267
Data Link Layer 176
Data Security 265
Data Terminal Equipment 90
Data Transmission 89, 90, 105, 108
Database Middleware 425
Decryption 239–245
Delivery 282
Demand Fluctuation Stock 293
Demilitarized Zone (DMZ) 224
Denial of Service 208
Deny All 208
Desktop Agents 258–259
Diffie-Hellman Key Exchange 321
Digital Certificate 257, 258
Digital Economy 45, 46
Digital Goods and Digitally Deliverable Services 298
Digital Products 53, 61
Digital Signature 259–260
Digital Signature Standard (DSS) 260
Disclosure 236
Disintermediation 5–8
Display Advertising 371
Distinguished Name 257
Distribution 282, 297, 299, 301, 302
Distribution Chain 7, 9
Distribution Channels 362
DNS 90, 127–131
DNS Spoofing 313
Document Object Model (DOM) 186–187
Domain Name System 90, 127–131, 213
Name Resolution 128, 129
Name Server 128, 129, 213
Name Space 127
Registering 130
Resolver 130
DTE 90
Dynamic HTML 186–187
E
eCash 309, 310–312
ECDSA 261
EDI 94
EDI Standards 70–75
ANSI ASC X12 71–73
EDIFACT 73–74
X.435 74
E-learning 37
Electronic Auctions 34
Electronic Banking 35
Electronic Checks
FSTC 324, 325
Mandate 325, 326
Netcheque 326
Electronic Commerce
Applications 34
Architecture 88, 89, 91
B2B 16–19
B2C 20–22
B2E 35
Benefits 5, 6
Business Models 46–47
C2B 22, 23
C2C 22, 24
Classification 15–16
Consumer’s Perspective 12
Definition 2
Elements 4–5
Framework 89
Impact 7–8
Industry Perspective 9, 10–11
Intra Organization 24–27
Learning 37
Risks 14
What is 1, 2
Electronic Community 8–10
Electronic Data Interchange (EDI) 63, 94, 281, 287, 290
Application layer 69
Architecture 68
Business forms 69
Data Transport Layer 75
Definition 67–68
Document Standards 70
Interconnection layer 63
Electronic Document Exchange 68–69
Electronic Learning 37
Electronic Mail 134, 138–142, 214, 260–261
Applications 142
Message Format 140, 141
Security 261
Electronic Manufacturing Service
Electronic Market 7–15
Electronic Payment Systems: see Payement Systems
Electronic Searching 36
Electronic Serial Number(ESN) 444
Electronic Store Model 58, 59
Electronic Trading 40
Elliptic Curve Algorithm 260
E-mail Marketing 371
Encryption 93, 209, 228, 236, 239, 247–248
Asymmetric 239, 241
Symmetric 239
Encryption Key 241, 246–248, 321
Enhanced Competition 286
E-Procurement 291
Ethernet 111–116
Extensible Markup Language 91, 92
Extranets 2
F
Fiber Optic Cable 107
MuliMode Step Index 108
MultiMode Graded Index 108
Single Mode Fiber 109
File Transfer Protocol (FTP) 118, 135–136
Firewalls 215
Application Level Gateway 219
Circuit Level Gateway 218
Limitations 222
Packet Filtering 216
Stateful Inspection 221
Web Server Placement 223, 224
First Virtual 249, 330–336
Framework of Electronic Commerce 88
Freeware Model 49
FSTC Electronic Check 324–326
FTPMAIL 143
G
Gateway 216
GIF 191–194
Globalization 286, 304
Global Packet Radio Service 438
SGSN 439
GGSN 439
packet-switching 439
Graphic Formats
GIF 191–194
JPEG 191–195
PNG 192–193
Raster 191
TIFF 192
Transparent
Vector 191–192
GSM 435
Base Station 436
Home Location Register 437
Mobile Station 436
Spectral Allocation 437
Visiting Location Register 437
GUI 139
H
Helper Applications 195–198
HEPNET 89
Hines 283
Hit Ratio 38
Host-to-Network Access Layer 118
HTML
Anchor Tag 164
Block Structuring Tags 160
Editors 187, 188
Form Tags 168, 169
Image Tag 164
List Tags 162
Text Formatting Tags 158
HTTP_Accept 146–176
HTTP_User_Agent 146–176
Hyper Text Transfer Protocol (HTTP) 91, 92, 144–149
HyperText Markup Language (HTML) 91–92, 154
I
IANA 89
IDEA 245
IEEE 112, 115
iHTML 449–454
iKP
Image Formats
GIF 190–194
JPEG 190–195
PNG 191–193
Raster 190
TIFF 191
Transparent
Vector 190–191
iMode 448
Impact of production planning and Inventory 281
Impact on distribution 281, 297
Impact on procurement 281, 290
Impression 400
Indian Customs EDI System (ICES) 83
ICES/Export 87–88
ICES/Import 84–86
Infomediary Model 51
Information Content Model 48
Information Distribution 90
Information Exchange Model 49
Information Filtering Agents 472
Information Management 282
Information Repository 90, 91
Information Technology Act 97–98
Integrity 237, 258
Intelligent Agents
Control 477
Cooperation 472
Coordination 475
Standards and Protocols 477
Integrated Marketing Communication (IMC) 369
Interactive Advertising 372
Intermediary 5, 7, 18, 53
Internet Access Provision 54
Internet Advertising Models 404
Banner Advertisements 404–408
Corporate Web-Sites 413
Customized Banner Advertisment 410
Interstitials 398, 415
Microsites 413
Opt-in’s 416
Push Broadcasting 414
Screensavers 414
Sponsoring Content 412– 413
Strength 401
Superstitials 412
Weakness 416
Internet Agents 472
Information Filtering Agents 472
Information Retrieval Agents 473
Notification agents 433, 473
Web Search Agents 472
Web Server Agents 472
Mobile 485
Internet Assigned Numbers Authority 89
Internet Banking 35
Internet Industry Structure 130, 131
Internet Infrastructure Attacks 205
Internet Layer 119, 120
Internet Protocol 89, 119, 120
Addressing 121, 122
Internet Service Provider (ISP) 53, 54, 131, 132
NAP 130, 131
PNAP 130, 131
POP 131
Interstitials 415, 416
Intranet 23–27, 473
Intranet Agents 472
International Mobile Equipment Identity 437
Inventory management 282, 287
Inventory Planning 359
IP Address 120–125
IP Spoofing 212–213
IPV6 89, 90
ISAPI 181–182
IT Objective 302
J
Java 37, 189, 424
Java Applets 189, 197, 198, 409– 410
JAVAScript 187, 189
JPEG 194, 195
Jscripts 189
K
Kasbah 480
Kerberos, 253–256
Authentication Server 252
Ticket Granting Server 253
Key Distribution 242, 251, 252
Key Length 244–245
Key Management 252, 257, 262
Knowledge Query and Manipulation Language (KQML) 474
L
LAN 88, 102–116
Location and Search Service 463
Long Tail Effects 352
M
Malicious Code 205
Mandate 325, 326
Mango Growers 283
Mango plantation 284
Manufacturing Model 47, 60
Manufacturing Planning 292
Marketing Communication 348, 369
Material Planning 359
Masquerade 236
MD5 248, 249, 258
Media Access Layer 110, 115
Media Access Unit 111–112
Meet-in-the-Middle attack, 244
Mentzer 283
Message Digest 5 (MD5) 248, 249, 258
Message Digest Algorithm 248, 249
Message Format 140, 141
Message Integrity 93, 246
Metamediary Model 55, 56
Metered Service Model 55, 56
MicroMint 314, 315
Microsites 413
MilliCent 314, 315
MIME 140–142, 145, 176, 195
MiniPay 216, 217
Mobile Agents 484
Mobile Auction 463
Mobile Commerce 426–431
benefits 427–430
definition 426–427
devices 426–427
framework 431
impediments 430–431
payment systems 430, 455–458
publishing languages 449–455
Security 454
Mobile Identification Number 434
Mobile Integrity Check Protocol 435
Mobile Payment Models 430, 457, 458
Acquirer Centric 458
Issuer Centric 458
Mobile N/W Operator Centric 458
Mobile Service Center 433
mod_perl 181–182
Mondex 311–312, 330–336
Multimedia Objects 195
Multi-vendor Catalog 56
N
Name Resolution 127–128, 129
Name Server 128, 129, 199
Name Space 127
NAP 130, 131
National Telecom Policy 96
NetBill 183, 330–336
Net-Buyers 401
NetCheque 326–327, 330–336
Net-Consumers 401
NetFare 317, 318
Net-Surfers 401
Network Access Point 130, 131
Network Address Translation 277
Network Infrastructure 89, 90
Network Layer 119
Network News Transfer Protocol (NNTP) 135
Network Topologies 102–106
Bus 103
Mixed 105
Ring 104
Star 104
Networks
ARPANET 89, 117
BITNET 89, 109
CSNET 89
HEPNET 89
LAN 89, 102-116
Packet Switched 89
SPAN 89, 118
WAN 117
Nonce 253
Non-Repudiation 237, 238, 258
Notification Agents 473, 483
NSAPI 156
O
OECD 96
One-time Key 262– 263
One-time Password 206, 210
Online Payment System: see Electronic Payment Systems,
Open Market 9
Operational Improvements 286
Opt-in Advertising Model 416
OSI 89, 117
Outsourcing 286, 304
Overproduction Stock 293
P
Packet Filtering Firewall 216–217
Packet Switched Network 89, 110
Packet Sniffer 205
Partner Collaboration 295
Payment Categories
Business Payments 309
Consumer Payments 309
Micro Payments 309
Payment Characteristics
Acceptability 308
Convertibility 308
Efficiency 308
Flexibility 308
Reliability 308
Scalability 308
Security 308–327
Usability 308
Payment Gateway 265–266, 267, 323
Payment System
CyberCash 318, 321, 330–336
CyberCoin 318
eCash 309, 330–336
First Virtual 328, 330–336
FSTC Electronic Check 324, 325
iKP 319–321
Mandate 325, 326
MicroMint 314, 329–336
MilliCent 312
MiniPay 316, 330–336
Mondex 311, 330–336
NetBill 315, 330–336
NetCheque 326, 330–336
NetFare 317
SET 322, 330–336
Perlscript 184, 185
Pesonal Digital Assistant 426
Personal Selling 368
PGP 261
Physical Distribution 297, 348, 358
Physical Goods 298, 300, 301
Physical Layer 118
Plaintext 240–248
PNAP 130–131
PNG 192, 193
Point of Presence (POP) 130, 131
Portable Access 24
Portal 50
Post Deployment 289
PPP 118, 119, 132
Pretty Good Privacy 261
Price 363
Pricing 348
Privacy 225, 229, 235, 238
Privacy Enhanced Mail 262, 263
Probe 152
Procurement 282, 290, 291
Product 348, 349
Production 282, 292, 294, 297, 305
Proliferation of E-Commerce 287
Promotion 348, 366
Promotions 399
Protocols
Address Resolution Protocol (ARP) 123, 210, 211
Agent Transfer Protocol (ATP) 478
ALOHA 111
CSMA 111–112
CSMA/CD 112
FTP 127, 135–138
HTTP 127, 144–148
Internet Protocol 118–124
Media Access 115
PPP 111, 118
SATP 477
Secure HTTP (SHTTP) 267–268, 263
SET 265–267, 322, 330–336
SHEN 268
Simple Message Transfer (SMTP) 127, 139–142
TCP/IP 89, 90, 117–130
Transmission Control Protocol 124, 125
User Datagram Protocol (UDP) 126
Wireless Access Protocol (WAP) 46
Proxy Server 206, 214, 215, 219–221
Public Key Algorithms 245–246
Public Key Cryptography 256
Public Key Cryptosystem 241
Public Key Infrastructure (PKI) 257, 258
Push Broadcasting Model 414
Q
Quality Assurance 282
R
Raster Images 190, 191
Reflection Attack 250–251, 252
Registration Authority 257
Reliability 308
Replay Attack 252
Resolver 129
Reversible Digital Signature Algortithm (rDSA) 260, 261
RFC 821 142
RFC 822 139, 140
RGB Color 189, 190
Root Compromise 152
Router 102, 106, 208–209, 217–218
RSA Algorithm 246–247, 273–275
S
Safety Stock 292
Sales Promotion 368
Scheduling 282
Screensavers Advertising Model 414
Secure Electronic Transaction (SET) 265–267, 322, 330–336
Secure Hash Algorithm (SHA) 249
Secure HTTP (SHTTP) 267, 268
Secure Socket Layer (SSL) 263–265
Security
Electronic Mail 261
Policy 205, 206
Services 213, 225
Site 207–208, 209
Transactions 265–267
Security Policy 205, 206
Security Practices 230, 231
Server Privileges 225
Server Side Includes (SSI) 181–185
Services Security 213–214, 225
Session Layer 218, 221
SET 265–267, 322, 330–336
SHEN 268
Simple Agent Transfer Protocol (SATP) 477
Simple Mail Transfer Protocol
(SMTP) 127, 139–142
Site Security 207, 208
SLIP 118, 119
Sniffing 208–210
SPAN 89, 118
Sponsored Content Model 412
Sponsoring Process 414
Spoofing
ARP 211, 212
DNS 213, 214
IP 211–213
Standard Generalized Markup Language (SGML) 157
Stateful Inspection Firewall 221
Subscription Model 50
Superstitials 416
Supply Chain Complexity 286
Supply Chain Management 39, 40, 66, 283, 286, 287, 305, 486
Supply Sensing 295
T
Tagged Information File Format (TIFF) 192
TCP/IP 89, 90, 117–130
Telescript 474
Telnet 126, 128
Temporal Key Integrity Protocol 455
Tete-a-tete 483
The Challenge 289
Timing Modification 236
Token Ring 118
Tool Command Language 475
Trading Process 64
Traffic Analysis 236
Transaction Security 235, 236
Authentication 93, 96, 97, 214, 237, 249–250
Authorization 238
Confidentiality 238, 266–267
Integrity 237, 258
Non Repudiation 237, 258, 262–263
Transmission Control Protocol, 89, 90, 123, 124–125
Transmission Media 105
CAT-3 106
CAT-5 106
Coaxial Cable 106
Fiber Optic 107
Infrared 110
Radio Frequency 110
Twisted Pair 114
Wireless 109
Transparent Image 193, 194
Transport Layer 124
Triple DES 244, 245
Trojan Horse 153, 214–215
True Image Formats 192
Trust Exploitation 134
Twisted Pair 114
U
Uniform Resource Locator (URL) 91, 92, 144–147
User-Agent 146–148
User Datagram Protocol (UDP) 126, 127
V
Value Added Network 67–68, 75–78
Providers 78
Value Chain 6–9
vBNS 131
VBScript 188, 189
Vector Images 190–191
Virtual Classroom 37, 38
Virtual Community 12–13
Virtual Corporation 12, 40, 41
Virtual Library 47
Virtual Manufacturing 17, 18
Virtual Shopping Agent 481
Virtual Supply Chain 16–17
Virus 93, 205, 275–279
Vortals 50
VRML 197–198
VSNL 54, 97
Vulnerability 201–204
Protocol 203
Technical 202
W
WAN 117
WAP 46
Weaknesses of Internet Advertising 416
Web Browser
Internet Explorer 156
Mosaic 155, 156
Netscape Navigator 155
Web Hosting & Internet Services 55
Web Image Formats 192
Web Search Agents 472
Web Server Agents 472
Web Server Security 224
CGIWrap 232
Disabling Features 227
File Permissions 226
Privileges 225
Server Logs 229
Web Servers
Apache 149, 151, 152
NCSA 149–151
Wide Area Networks 117
Wireless Access Protocol 46, 443
Wireless Application Environment 444–446
Wireless Datagram Protocol 444
Wireless Networks 431
AMPS 425, 426
CDMA 432, 434
CDMA2000 432
EDGE 426, 438
GPRS 426, 438
GSM 425, 435
TD-CDMA 441
WCDMA 440–441
Wireless Session Protocol 444–446
Wireless Transaction Protocol 444–446
Wireless Transport Layer Security 444–446
Wireless Transmission 109
Infrared Based 110
Radio-Based 110
World Wide Web (WWW) 143–144
Server 148–150
WML 449-450
X
X.25 118
X.435 67
X.500 Distinguished Name
X.509 257, 258
XML 92–93